var/home/core/zuul-output/0000755000175000017500000000000015111302664014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111316531015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005310162015111316522017670 0ustar rootrootNov 25 10:36:41 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 10:36:42 crc restorecon[4586]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:42 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:36:43 crc restorecon[4586]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 10:36:43 crc kubenswrapper[4696]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:36:43 crc kubenswrapper[4696]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 10:36:43 crc kubenswrapper[4696]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:36:43 crc kubenswrapper[4696]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:36:43 crc kubenswrapper[4696]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 10:36:43 crc kubenswrapper[4696]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.671268 4696 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677893 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677934 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677943 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677953 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677962 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677970 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677980 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.677992 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678003 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678014 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678033 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678044 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678053 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678063 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678075 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678086 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678097 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678108 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678119 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678130 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678141 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678152 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678162 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678174 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678185 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678195 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678204 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678214 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678221 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678232 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678243 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678251 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678259 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678268 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678277 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678285 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678293 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678303 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678312 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678320 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678328 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678336 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678344 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678352 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678359 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678367 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678378 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678387 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678395 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678403 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678411 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678420 4696 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678428 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678436 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678443 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678451 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678459 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678467 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678477 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678486 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678495 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678503 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678513 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678521 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678529 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678537 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678548 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678558 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678567 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678575 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.678583 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681852 4696 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681885 4696 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681897 4696 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681904 4696 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681911 4696 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681916 4696 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681922 4696 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681928 4696 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681933 4696 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681937 4696 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681942 4696 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681946 4696 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681951 4696 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681955 4696 flags.go:64] FLAG: --cgroup-root="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681959 4696 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681963 4696 flags.go:64] FLAG: --client-ca-file="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681967 4696 flags.go:64] FLAG: --cloud-config="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681971 4696 flags.go:64] FLAG: --cloud-provider="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681976 4696 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681982 4696 flags.go:64] FLAG: --cluster-domain="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681986 4696 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681990 4696 flags.go:64] FLAG: --config-dir="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.681995 4696 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682000 4696 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682006 4696 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682010 4696 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682015 4696 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682021 4696 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682026 4696 flags.go:64] FLAG: --contention-profiling="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682030 4696 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682035 4696 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682040 4696 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682044 4696 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682051 4696 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682055 4696 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682061 4696 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682070 4696 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682079 4696 flags.go:64] FLAG: --enable-server="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682087 4696 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682095 4696 flags.go:64] FLAG: --event-burst="100" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682101 4696 flags.go:64] FLAG: --event-qps="50" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682106 4696 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682112 4696 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682117 4696 flags.go:64] FLAG: --eviction-hard="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682124 4696 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682129 4696 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682134 4696 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682139 4696 flags.go:64] FLAG: --eviction-soft="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682145 4696 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682149 4696 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682153 4696 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682157 4696 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682161 4696 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682165 4696 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682172 4696 flags.go:64] FLAG: --feature-gates="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682178 4696 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682183 4696 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682188 4696 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682193 4696 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682197 4696 flags.go:64] FLAG: --healthz-port="10248" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682202 4696 flags.go:64] FLAG: --help="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682206 4696 flags.go:64] FLAG: --hostname-override="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682210 4696 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682214 4696 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682218 4696 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682222 4696 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682226 4696 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682230 4696 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682234 4696 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682238 4696 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682242 4696 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682247 4696 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682251 4696 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682255 4696 flags.go:64] FLAG: --kube-reserved="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682259 4696 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682263 4696 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682267 4696 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682271 4696 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682276 4696 flags.go:64] FLAG: --lock-file="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682280 4696 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682285 4696 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682290 4696 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682306 4696 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682313 4696 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682318 4696 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682323 4696 flags.go:64] FLAG: --logging-format="text" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682327 4696 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682332 4696 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682337 4696 flags.go:64] FLAG: --manifest-url="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682341 4696 flags.go:64] FLAG: --manifest-url-header="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682348 4696 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682353 4696 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682359 4696 flags.go:64] FLAG: --max-pods="110" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682363 4696 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682367 4696 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682371 4696 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682375 4696 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682379 4696 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682384 4696 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682388 4696 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682401 4696 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682406 4696 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682410 4696 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682414 4696 flags.go:64] FLAG: --pod-cidr="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682419 4696 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682425 4696 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682429 4696 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682433 4696 flags.go:64] FLAG: --pods-per-core="0" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682438 4696 flags.go:64] FLAG: --port="10250" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682442 4696 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682447 4696 flags.go:64] FLAG: --provider-id="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682451 4696 flags.go:64] FLAG: --qos-reserved="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682455 4696 flags.go:64] FLAG: --read-only-port="10255" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682459 4696 flags.go:64] FLAG: --register-node="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682463 4696 flags.go:64] FLAG: --register-schedulable="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682467 4696 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682475 4696 flags.go:64] FLAG: --registry-burst="10" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682479 4696 flags.go:64] FLAG: --registry-qps="5" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682483 4696 flags.go:64] FLAG: --reserved-cpus="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682487 4696 flags.go:64] FLAG: --reserved-memory="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682493 4696 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682497 4696 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682501 4696 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682505 4696 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682509 4696 flags.go:64] FLAG: --runonce="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682514 4696 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682519 4696 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682523 4696 flags.go:64] FLAG: --seccomp-default="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682527 4696 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682531 4696 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682536 4696 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682540 4696 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682544 4696 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682549 4696 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682553 4696 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682558 4696 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682562 4696 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682567 4696 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682571 4696 flags.go:64] FLAG: --system-cgroups="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682575 4696 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682581 4696 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682585 4696 flags.go:64] FLAG: --tls-cert-file="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682590 4696 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682594 4696 flags.go:64] FLAG: --tls-min-version="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682598 4696 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682603 4696 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682607 4696 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682611 4696 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682615 4696 flags.go:64] FLAG: --v="2" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682620 4696 flags.go:64] FLAG: --version="false" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682627 4696 flags.go:64] FLAG: --vmodule="" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682632 4696 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.682636 4696 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682782 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682790 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682795 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682800 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682806 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682809 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682813 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682817 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682822 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682827 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682831 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682834 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682839 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682843 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682846 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682850 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682853 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682857 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682861 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682865 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682869 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682873 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682877 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682882 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682886 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682889 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682893 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682897 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682900 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682904 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682908 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682912 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682916 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682920 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682924 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682928 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682933 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682937 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682940 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682945 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682948 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682952 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682955 4696 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682959 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682963 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682967 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682971 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682976 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682979 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682983 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682986 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682990 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682993 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.682996 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683000 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683003 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683007 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683010 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683013 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683017 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683020 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683025 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683031 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683068 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683076 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683081 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683086 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683091 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683095 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683099 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.683102 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.683116 4696 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.705814 4696 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.705862 4696 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705947 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705956 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705960 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705965 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705969 4696 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705973 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705976 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705981 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705988 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705994 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.705998 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706002 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706006 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706010 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706014 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706018 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706023 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706026 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706030 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706033 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706037 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706040 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706043 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706048 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706051 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706055 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706059 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706064 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706069 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706074 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706083 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706090 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706095 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706099 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706104 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706108 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706112 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706116 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706121 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706125 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706129 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706133 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706136 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706140 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706143 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706147 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706150 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706154 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706157 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706160 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706163 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706167 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706170 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706174 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706177 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706180 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706185 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706188 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706192 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706196 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706200 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706204 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706208 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706213 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706219 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706224 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706228 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706232 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706237 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706242 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706246 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.706252 4696 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706367 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706373 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706378 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706382 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706385 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706389 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706393 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706397 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706402 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706406 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706410 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706414 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706418 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706422 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706425 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706428 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706432 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706436 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706440 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706444 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706448 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706452 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706457 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706461 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706464 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706468 4696 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706472 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706476 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706479 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706482 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706486 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706489 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706493 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706496 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706500 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706503 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706507 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706512 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706516 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706519 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706522 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706526 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706529 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706533 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706537 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706540 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706543 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706547 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706550 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706553 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706557 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706560 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706563 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706567 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706572 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706575 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706579 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706582 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706586 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706589 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706592 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706596 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706599 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706603 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706606 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706609 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706613 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706616 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706619 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706623 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.706627 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.706632 4696 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.709136 4696 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.720485 4696 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.720581 4696 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.722260 4696 server.go:997] "Starting client certificate rotation" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.722284 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.722566 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-10 18:45:44.747436417 +0000 UTC Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.722649 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 368h9m1.024790293s for next certificate rotation Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.859533 4696 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.861340 4696 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.894425 4696 log.go:25] "Validated CRI v1 runtime API" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.927188 4696 log.go:25] "Validated CRI v1 image API" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.928536 4696 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.934952 4696 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-10-30-25-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.934992 4696 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.957417 4696 manager.go:217] Machine: {Timestamp:2025-11-25 10:36:43.954234341 +0000 UTC m=+0.767851008 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:25dbb5c2-d084-42b3-95e0-6e02dc598ddc BootID:8849372d-5f88-4267-ab96-46be61e7678f Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:a5:97:b8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:a5:97:b8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:3a:7b:9c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:66:44:75 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:68:3f:8c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f4:dc:77 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8a:bb:16:36:53:bc Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:62:50:96:a2:c2:79 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.957778 4696 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.957969 4696 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.960474 4696 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.960856 4696 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.960915 4696 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.961215 4696 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.961232 4696 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.961921 4696 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.961970 4696 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.962206 4696 state_mem.go:36] "Initialized new in-memory state store" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.962330 4696 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.967246 4696 kubelet.go:418] "Attempting to sync node with API server" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.967287 4696 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.967333 4696 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.967358 4696 kubelet.go:324] "Adding apiserver pod source" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.967387 4696 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.975397 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:43 crc kubenswrapper[4696]: E1125 10:36:43.975454 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.975558 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:43 crc kubenswrapper[4696]: E1125 10:36:43.975688 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.977510 4696 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.978797 4696 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.981654 4696 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983142 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983166 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983176 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983184 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983196 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983204 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983212 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983224 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983234 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983244 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983273 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.983282 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.986826 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.987254 4696 server.go:1280] "Started kubelet" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.988384 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.988756 4696 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.988874 4696 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 10:36:43 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.989754 4696 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990373 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990407 4696 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990464 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 01:12:51.305109327 +0000 UTC Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990548 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 902h36m7.314564777s for next certificate rotation Nov 25 10:36:43 crc kubenswrapper[4696]: E1125 10:36:43.990612 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990669 4696 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990704 4696 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.990846 4696 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 10:36:43 crc kubenswrapper[4696]: E1125 10:36:43.991227 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="200ms" Nov 25 10:36:43 crc kubenswrapper[4696]: W1125 10:36:43.991805 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:43 crc kubenswrapper[4696]: E1125 10:36:43.991930 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.994188 4696 server.go:460] "Adding debug handlers to kubelet server" Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.994652 4696 factory.go:55] Registering systemd factory Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.994700 4696 factory.go:221] Registration of the systemd container factory successfully Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.995958 4696 factory.go:153] Registering CRI-O factory Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.995997 4696 factory.go:221] Registration of the crio container factory successfully Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.996077 4696 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.996100 4696 factory.go:103] Registering Raw factory Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.996118 4696 manager.go:1196] Started watching for new ooms in manager Nov 25 10:36:43 crc kubenswrapper[4696]: I1125 10:36:43.996809 4696 manager.go:319] Starting recovery of all containers Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:43.996570 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.38:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b3997fca565c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:36:43.987224001 +0000 UTC m=+0.800840608,LastTimestamp:2025-11-25 10:36:43.987224001 +0000 UTC m=+0.800840608,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008785 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008835 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008851 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008866 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008881 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008894 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008906 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008919 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008934 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008946 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008960 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008974 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.008987 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009006 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009021 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009033 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009046 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009059 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009074 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009088 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009100 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009114 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009126 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009139 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009152 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009165 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009182 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009196 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009210 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009225 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009240 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009253 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009265 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009278 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009290 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009303 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009316 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009328 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009341 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009354 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009366 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009381 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009394 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009407 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009420 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009435 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009449 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009463 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009476 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009491 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009504 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009517 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009540 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009556 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009572 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009587 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009601 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009622 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009636 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009649 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009667 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009748 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009764 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009777 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009790 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009804 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009818 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009831 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009844 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009857 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009869 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009882 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009896 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009908 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009921 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009933 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009946 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009962 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009975 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.009989 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010003 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010015 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010028 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010042 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010083 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010095 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010108 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010137 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010150 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010163 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010176 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010188 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010200 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010214 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010228 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010240 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010252 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010264 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010276 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010289 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010301 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010315 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010328 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010340 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010367 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010382 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010396 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010410 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010424 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010451 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010466 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010481 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010495 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010511 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010523 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010539 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010552 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010565 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010579 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010591 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010604 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010617 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010630 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010644 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010657 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010693 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010707 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010722 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010735 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010748 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010762 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010776 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010788 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010800 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010812 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010825 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010838 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010852 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010864 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010877 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010890 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010905 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010919 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010933 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010946 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010961 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010973 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010985 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.010997 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011011 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011024 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011036 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011051 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011065 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011077 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011091 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011106 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011120 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011134 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011147 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011159 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011172 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011185 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011198 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011211 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.011226 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.014956 4696 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.014994 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015012 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015026 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015038 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015053 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015070 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015085 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015098 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015111 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015126 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015151 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015163 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015176 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015189 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015203 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015216 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015230 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015245 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015259 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015274 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015315 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015329 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015352 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015365 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015379 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015395 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015410 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015425 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015439 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015452 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015468 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015481 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015495 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015509 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015525 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015538 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015551 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015566 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015581 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015595 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015610 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015622 4696 reconstruct.go:97] "Volume reconstruction finished" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.015630 4696 reconciler.go:26] "Reconciler: start to sync state" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.024042 4696 manager.go:324] Recovery completed Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.034155 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.035356 4696 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.035648 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.035702 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.035711 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.036556 4696 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.036581 4696 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.036606 4696 state_mem.go:36] "Initialized new in-memory state store" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.040927 4696 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.040964 4696 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.040989 4696 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.041024 4696 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.041840 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.042002 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.060044 4696 policy_none.go:49] "None policy: Start" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.060934 4696 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.060972 4696 state_mem.go:35] "Initializing new in-memory state store" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.091319 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.115838 4696 manager.go:334] "Starting Device Plugin manager" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.116113 4696 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.116126 4696 server.go:79] "Starting device plugin registration server" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.117121 4696 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.117158 4696 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.117751 4696 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.117843 4696 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.117852 4696 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.124256 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.141570 4696 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.141679 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.143408 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.143455 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.143468 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.143637 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.143867 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.143928 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.144747 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.144780 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.144792 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.144917 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.146486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.146531 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.146542 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.149087 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.149156 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.150501 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.150535 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.150565 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.150735 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.150844 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.150895 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152543 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152580 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152554 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152693 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152706 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152721 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152775 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152788 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.152858 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.153088 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.153157 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.153583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.153610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.153625 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.153987 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.154048 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.154622 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.154656 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.154685 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.156933 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.156982 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.156995 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.192429 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="400ms" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217243 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217285 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217311 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217334 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217359 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217379 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217401 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217404 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217469 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217538 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217559 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217579 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217597 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217612 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217644 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.217695 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.218766 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.218809 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.218820 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.218866 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.219426 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319124 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319188 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319228 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319295 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319325 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319324 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319354 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319349 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319414 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319430 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319439 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319392 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319458 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319500 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319508 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319526 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319555 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319555 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319579 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319587 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319611 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319619 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319640 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319643 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319678 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319688 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319712 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.319782 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.320146 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.419857 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.421000 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.421027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.421035 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.421056 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.421294 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.478755 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.485320 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.504758 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.521472 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.523018 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-13a4e55c0b3ec3468831ae233f4421c434e542cdb87b8b75751174e3931b308f WatchSource:0}: Error finding container 13a4e55c0b3ec3468831ae233f4421c434e542cdb87b8b75751174e3931b308f: Status 404 returned error can't find the container with id 13a4e55c0b3ec3468831ae233f4421c434e542cdb87b8b75751174e3931b308f Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.523304 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-3c5fca2e5ed769e5d185d84be64255d91d266873293e02cb547468a09beb0b74 WatchSource:0}: Error finding container 3c5fca2e5ed769e5d185d84be64255d91d266873293e02cb547468a09beb0b74: Status 404 returned error can't find the container with id 3c5fca2e5ed769e5d185d84be64255d91d266873293e02cb547468a09beb0b74 Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.525603 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.526621 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ee23788b5473cb22bf2a6147e6cef1e2f2dce63b0e86dc2d8661fc353f3847c3 WatchSource:0}: Error finding container ee23788b5473cb22bf2a6147e6cef1e2f2dce63b0e86dc2d8661fc353f3847c3: Status 404 returned error can't find the container with id ee23788b5473cb22bf2a6147e6cef1e2f2dce63b0e86dc2d8661fc353f3847c3 Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.534709 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9b790c952d51f5015e48435a74bf9a593f6c50b80d807aafc92f696c80d64259 WatchSource:0}: Error finding container 9b790c952d51f5015e48435a74bf9a593f6c50b80d807aafc92f696c80d64259: Status 404 returned error can't find the container with id 9b790c952d51f5015e48435a74bf9a593f6c50b80d807aafc92f696c80d64259 Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.546885 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-d248aee15bbaf6605862afdbf1302159ccd0c08489522cbc09371f6bdcad00b0 WatchSource:0}: Error finding container d248aee15bbaf6605862afdbf1302159ccd0c08489522cbc09371f6bdcad00b0: Status 404 returned error can't find the container with id d248aee15bbaf6605862afdbf1302159ccd0c08489522cbc09371f6bdcad00b0 Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.593387 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="800ms" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.821595 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.822988 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.823033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.823045 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.823078 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.823546 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Nov 25 10:36:44 crc kubenswrapper[4696]: W1125 10:36:44.849549 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:44 crc kubenswrapper[4696]: E1125 10:36:44.849640 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:44 crc kubenswrapper[4696]: I1125 10:36:44.988946 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:45 crc kubenswrapper[4696]: W1125 10:36:45.042171 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:45 crc kubenswrapper[4696]: E1125 10:36:45.042250 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.044893 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"13a4e55c0b3ec3468831ae233f4421c434e542cdb87b8b75751174e3931b308f"} Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.045691 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3c5fca2e5ed769e5d185d84be64255d91d266873293e02cb547468a09beb0b74"} Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.046471 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d248aee15bbaf6605862afdbf1302159ccd0c08489522cbc09371f6bdcad00b0"} Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.047932 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9b790c952d51f5015e48435a74bf9a593f6c50b80d807aafc92f696c80d64259"} Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.049039 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ee23788b5473cb22bf2a6147e6cef1e2f2dce63b0e86dc2d8661fc353f3847c3"} Nov 25 10:36:45 crc kubenswrapper[4696]: W1125 10:36:45.262053 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:45 crc kubenswrapper[4696]: E1125 10:36:45.262224 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:45 crc kubenswrapper[4696]: W1125 10:36:45.275259 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:45 crc kubenswrapper[4696]: E1125 10:36:45.275339 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:45 crc kubenswrapper[4696]: E1125 10:36:45.395248 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="1.6s" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.623814 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.625141 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.625177 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.625187 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.625210 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:36:45 crc kubenswrapper[4696]: E1125 10:36:45.625683 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Nov 25 10:36:45 crc kubenswrapper[4696]: I1125 10:36:45.989402 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.053318 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e" exitCode=0 Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.053413 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.053457 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.054340 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.054371 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.054383 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.054736 4696 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf" exitCode=0 Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.054801 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.054898 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.055898 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.055916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.055924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.056580 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.057586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.057616 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.057629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.058207 4696 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1" exitCode=0 Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.058268 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.058331 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.067835 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.067859 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.067867 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.080251 4696 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ebe196bdfbebf41e6668bcdcf12d335bb94611db1591d1b0183935f9b6d15704" exitCode=0 Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.080353 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ebe196bdfbebf41e6668bcdcf12d335bb94611db1591d1b0183935f9b6d15704"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.080418 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.081382 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.081405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.081412 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094184 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094228 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094234 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094239 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094248 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523"} Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094897 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.094946 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:46 crc kubenswrapper[4696]: W1125 10:36:46.804276 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:46 crc kubenswrapper[4696]: E1125 10:36:46.804424 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.38:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:36:46 crc kubenswrapper[4696]: I1125 10:36:46.989709 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.38:6443: connect: connection refused Nov 25 10:36:46 crc kubenswrapper[4696]: E1125 10:36:46.996716 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="3.2s" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.098496 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.098547 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.098564 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.099003 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.099912 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7264441bd8ecb4f8941bdca823785efd2ea998fbb88ef49f32577e274aff5a1a"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.099933 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.100256 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.100287 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.100300 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.100928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.101063 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.101161 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.101959 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.101991 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.102002 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.102013 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.103755 4696 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001" exitCode=0 Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.103818 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001"} Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.103957 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.104109 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.104874 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.104900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.104910 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.105003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.105024 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.105032 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.144496 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.226185 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.227797 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.227831 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.227839 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:47 crc kubenswrapper[4696]: I1125 10:36:47.227857 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:36:47 crc kubenswrapper[4696]: E1125 10:36:47.228563 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.38:6443: connect: connection refused" node="crc" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.109056 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.111404 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545" exitCode=255 Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.111473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545"} Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.111598 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.112559 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.112604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.112618 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.113286 4696 scope.go:117] "RemoveContainer" containerID="efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116357 4696 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280" exitCode=0 Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116408 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280"} Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116465 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116491 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116465 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116549 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.116643 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117581 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117748 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117757 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.117999 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.118037 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.118054 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.118040 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.118082 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.118062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:48 crc kubenswrapper[4696]: I1125 10:36:48.959236 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.122134 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.124600 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f"} Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.124752 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.126082 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.126126 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.126138 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129148 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6"} Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129221 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b"} Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129245 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5"} Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129257 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e"} Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129264 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e"} Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.129505 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.130507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.130547 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.130559 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.130871 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.130957 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.131030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.616069 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:49 crc kubenswrapper[4696]: I1125 10:36:49.794115 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.128188 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.131309 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.131425 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.132624 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.132738 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.132768 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.133424 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.133484 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.133501 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.429129 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.430881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.430932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.430949 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:50 crc kubenswrapper[4696]: I1125 10:36:50.430981 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.133463 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.133585 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.134310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.134345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.134359 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.135131 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.135175 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.135188 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.195499 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.195833 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.197249 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.197296 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.197316 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:51 crc kubenswrapper[4696]: I1125 10:36:51.206651 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:52 crc kubenswrapper[4696]: I1125 10:36:52.136104 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:52 crc kubenswrapper[4696]: I1125 10:36:52.136954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:52 crc kubenswrapper[4696]: I1125 10:36:52.136994 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:52 crc kubenswrapper[4696]: I1125 10:36:52.137007 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:53 crc kubenswrapper[4696]: I1125 10:36:53.143289 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:53 crc kubenswrapper[4696]: I1125 10:36:53.143565 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:53 crc kubenswrapper[4696]: I1125 10:36:53.144477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:53 crc kubenswrapper[4696]: I1125 10:36:53.144514 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:53 crc kubenswrapper[4696]: I1125 10:36:53.144524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:54 crc kubenswrapper[4696]: E1125 10:36:54.124339 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 10:36:54 crc kubenswrapper[4696]: I1125 10:36:54.889127 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:54 crc kubenswrapper[4696]: I1125 10:36:54.889320 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:54 crc kubenswrapper[4696]: I1125 10:36:54.890371 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:54 crc kubenswrapper[4696]: I1125 10:36:54.890423 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:54 crc kubenswrapper[4696]: I1125 10:36:54.890441 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:56 crc kubenswrapper[4696]: I1125 10:36:56.342180 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 10:36:56 crc kubenswrapper[4696]: I1125 10:36:56.342363 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:56 crc kubenswrapper[4696]: I1125 10:36:56.343352 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:56 crc kubenswrapper[4696]: I1125 10:36:56.343406 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:56 crc kubenswrapper[4696]: I1125 10:36:56.343418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.149274 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.149374 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.150093 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.150134 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.150146 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:36:57 crc kubenswrapper[4696]: W1125 10:36:57.818508 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.818630 4696 trace.go:236] Trace[1785468995]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:36:47.817) (total time: 10001ms): Nov 25 10:36:57 crc kubenswrapper[4696]: Trace[1785468995]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:36:57.818) Nov 25 10:36:57 crc kubenswrapper[4696]: Trace[1785468995]: [10.001534904s] [10.001534904s] END Nov 25 10:36:57 crc kubenswrapper[4696]: E1125 10:36:57.818692 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.889987 4696 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.890062 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 10:36:57 crc kubenswrapper[4696]: I1125 10:36:57.989910 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 10:36:58 crc kubenswrapper[4696]: I1125 10:36:58.067428 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 10:36:58 crc kubenswrapper[4696]: I1125 10:36:58.067510 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 10:36:58 crc kubenswrapper[4696]: I1125 10:36:58.073963 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 10:36:58 crc kubenswrapper[4696]: I1125 10:36:58.074126 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 10:36:58 crc kubenswrapper[4696]: I1125 10:36:58.969212 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]log ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]etcd ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/priority-and-fairness-filter ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-apiextensions-informers ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-apiextensions-controllers ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/crd-informer-synced ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-system-namespaces-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 25 10:36:58 crc kubenswrapper[4696]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 25 10:36:58 crc kubenswrapper[4696]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/bootstrap-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/start-kube-aggregator-informers ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-registration-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-discovery-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]autoregister-completion ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-openapi-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 25 10:36:58 crc kubenswrapper[4696]: livez check failed Nov 25 10:36:58 crc kubenswrapper[4696]: I1125 10:36:58.969294 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:36:59 crc kubenswrapper[4696]: I1125 10:36:59.617173 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 10:36:59 crc kubenswrapper[4696]: I1125 10:36:59.617259 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.069711 4696 trace.go:236] Trace[1145494285]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:36:51.387) (total time: 11682ms): Nov 25 10:37:03 crc kubenswrapper[4696]: Trace[1145494285]: ---"Objects listed" error: 11682ms (10:37:03.069) Nov 25 10:37:03 crc kubenswrapper[4696]: Trace[1145494285]: [11.682624787s] [11.682624787s] END Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.069744 4696 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 10:37:03 crc kubenswrapper[4696]: E1125 10:37:03.069781 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.071737 4696 trace.go:236] Trace[1795637937]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:36:48.190) (total time: 14881ms): Nov 25 10:37:03 crc kubenswrapper[4696]: Trace[1795637937]: ---"Objects listed" error: 14881ms (10:37:03.071) Nov 25 10:37:03 crc kubenswrapper[4696]: Trace[1795637937]: [14.881593926s] [14.881593926s] END Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.071774 4696 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.071924 4696 trace.go:236] Trace[1644536261]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:36:48.206) (total time: 14864ms): Nov 25 10:37:03 crc kubenswrapper[4696]: Trace[1644536261]: ---"Objects listed" error: 14864ms (10:37:03.071) Nov 25 10:37:03 crc kubenswrapper[4696]: Trace[1644536261]: [14.864996647s] [14.864996647s] END Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.071951 4696 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 10:37:03 crc kubenswrapper[4696]: E1125 10:37:03.073118 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.073239 4696 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.875707 4696 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.965178 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.967167 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.967375 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.971763 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.977928 4696 apiserver.go:52] "Watching apiserver" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.982531 4696 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.982860 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-vfbbz","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-l8nb9","openshift-machine-config-operator/machine-config-daemon-2mdzn","openshift-multus/multus-additional-cni-plugins-fb7d4","openshift-multus/multus-g77vb","openshift-network-diagnostics/network-check-target-xd92c"] Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983155 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983210 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983234 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:03 crc kubenswrapper[4696]: E1125 10:37:03.983268 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983452 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983477 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:03 crc kubenswrapper[4696]: E1125 10:37:03.983502 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983765 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983833 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:03 crc kubenswrapper[4696]: E1125 10:37:03.983840 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.983994 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.984036 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.984078 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.984303 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g77vb" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.988753 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.988825 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.988879 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.988945 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.989023 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.989084 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.989243 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.989766 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.989814 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.990050 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.990419 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.990579 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.990800 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.991922 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.991983 4696 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.991981 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992092 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992099 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992132 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992240 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992002 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992029 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992376 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992087 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992447 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992447 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992582 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992803 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992827 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992842 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.992948 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 10:37:03 crc kubenswrapper[4696]: I1125 10:37:03.993032 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.015196 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.025210 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.033789 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.037447 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.044793 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.059902 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.071358 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080185 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080230 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080251 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080286 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080302 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080318 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080335 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080349 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080370 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080386 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080401 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080419 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080456 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080471 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080488 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080517 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080533 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080550 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080566 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080581 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080599 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080603 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080614 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080613 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080681 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080701 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080718 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080733 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080750 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080768 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080784 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080790 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080816 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080831 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080847 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080863 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080888 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080908 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080926 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080943 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080959 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080974 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.080991 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081005 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081019 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081036 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081051 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081065 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081070 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081108 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081133 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081157 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081179 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081209 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081235 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081248 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081261 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081267 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081284 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081333 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081360 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081384 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081407 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081429 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081454 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081502 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081525 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081547 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081569 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081591 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081615 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081636 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081657 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081701 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081722 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081745 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081768 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081790 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081818 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081843 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081865 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081888 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081911 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082012 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082038 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082061 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082084 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082107 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082131 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082155 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082176 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082205 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082231 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082254 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082277 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082300 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082323 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082347 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082369 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082394 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082421 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082443 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082466 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082489 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082511 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082534 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082558 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082580 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082601 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082623 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082704 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082730 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082753 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082778 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082800 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082822 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082845 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082867 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082890 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082914 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082937 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082970 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082995 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083018 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083040 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083066 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083116 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083141 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083163 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083188 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083212 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083234 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083258 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083280 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083304 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083327 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083353 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083374 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083397 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083422 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083444 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083467 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083489 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083511 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083531 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083556 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083580 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083600 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083623 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083644 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083687 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083711 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083737 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083757 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083781 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083803 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083827 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083848 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083869 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083890 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083913 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083938 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083961 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083984 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084007 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084028 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084049 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084070 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084092 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084114 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084152 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084175 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084199 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084224 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084245 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084260 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084276 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084296 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084312 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084327 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084345 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084363 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084378 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084395 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084411 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084430 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084446 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084462 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084478 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084494 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084512 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084529 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084545 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084561 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084579 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084655 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084937 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084961 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084984 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085005 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085075 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67pdx\" (UniqueName: \"kubernetes.io/projected/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-kube-api-access-67pdx\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085110 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085136 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085155 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-netd\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085176 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-config\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085229 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085256 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085274 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085300 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085326 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085349 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-systemd-units\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085371 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-log-socket\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085392 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-system-cni-dir\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085414 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-hostroot\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085438 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-kubelet\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085484 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkcm8\" (UniqueName: \"kubernetes.io/projected/f2a63620-dc3c-467f-9cf5-d688f0e233c3-kube-api-access-wkcm8\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085509 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-mcd-auth-proxy-config\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085528 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-cnibin\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085550 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-cni-multus\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085571 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-ovn\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085594 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-k8s-cni-cncf-io\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085617 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-etc-kubernetes\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085639 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-netns\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085675 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-env-overrides\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085704 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcjxt\" (UniqueName: \"kubernetes.io/projected/a4cbf858-76fc-426e-8561-f898927b3bf5-kube-api-access-hcjxt\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-cnibin\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085760 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085782 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-system-cni-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085809 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085831 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-slash\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085856 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-multus-certs\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085918 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-proxy-tls\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085943 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-systemd\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085985 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-os-release\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086007 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-cni-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086027 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bcb29cb9-408c-48c2-bdca-2c3dce47b991-cni-binary-copy\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086042 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-etc-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086057 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086073 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-kubelet\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086087 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-daemon-config\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086105 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-var-lib-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086123 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-script-lib\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086142 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086159 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-netns\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086175 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-cni-bin\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086193 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3b9b7335-d993-46d0-a36a-7821280b7a91-hosts-file\") pod \"node-resolver-vfbbz\" (UID: \"3b9b7335-d993-46d0-a36a-7821280b7a91\") " pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086214 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpqmf\" (UniqueName: \"kubernetes.io/projected/3b9b7335-d993-46d0-a36a-7821280b7a91-kube-api-access-qpqmf\") pod \"node-resolver-vfbbz\" (UID: \"3b9b7335-d993-46d0-a36a-7821280b7a91\") " pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086261 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-os-release\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086290 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-conf-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086315 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086341 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086358 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086375 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-node-log\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086391 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-bin\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086410 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086428 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086444 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovn-node-metrics-cert\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a4cbf858-76fc-426e-8561-f898927b3bf5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086483 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-rootfs\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086503 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086524 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a4cbf858-76fc-426e-8561-f898927b3bf5-cni-binary-copy\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086549 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-socket-dir-parent\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5fkq\" (UniqueName: \"kubernetes.io/projected/bcb29cb9-408c-48c2-bdca-2c3dce47b991-kube-api-access-t5fkq\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086640 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086680 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086696 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086712 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.086727 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.090436 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.092849 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.094698 4696 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081348 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081436 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081440 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081612 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.081735 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082161 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082214 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082328 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082364 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082491 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082575 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082699 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.082941 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083432 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083718 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.083911 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084111 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084132 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084197 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084216 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084461 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084684 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.084870 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085517 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.085568 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.087818 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.088261 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.088476 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.088632 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.088690 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.088985 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.089386 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.090438 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.090556 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.090865 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.090968 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.091223 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.091655 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.092048 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.092263 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.092597 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.092741 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.093090 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.093540 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.094175 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.103777 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:04.603739407 +0000 UTC m=+21.417356064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.090072 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.095346 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.095356 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.095385 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.096261 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.096526 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.096904 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.097235 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.097381 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.099000 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.099914 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.100337 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.101660 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.103215 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.104382 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.104580 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.104748 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.105052 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.105341 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.105728 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.105940 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.106146 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.106706 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.106776 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.106937 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.106950 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.107317 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.107327 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.107413 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.107557 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.107574 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.107795 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.108022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.108260 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.108431 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.108500 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.108985 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.109423 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.109635 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.109852 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.116048 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.116258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.116470 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.116764 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.117263 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.117407 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.118064 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.119205 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.119242 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.121181 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.121378 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.121593 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.121812 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.121915 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.122032 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.122237 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.122955 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.125108 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.125264 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:37:04.625234699 +0000 UTC m=+21.438851286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.125724 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.126338 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.126610 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.127266 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.128446 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.128970 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.129180 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.129502 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.129773 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.129833 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.130093 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.130322 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.130400 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.130639 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.130757 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.130968 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.131196 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.131205 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.131390 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.131573 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.131603 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.131908 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.132046 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.132230 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.132319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.132861 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.133181 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.133453 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.133785 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.133887 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.135400 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.136183 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.136258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.146295 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.139546 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.136749 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.136805 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.136986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137012 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137104 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137109 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.143866 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137141 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137350 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.146481 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.139861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137366 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.137597 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.137910 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.138212 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.138403 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.138458 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.139923 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.139963 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.140097 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.140285 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.143948 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.144498 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.144582 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.144849 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.145060 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.145132 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.145147 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.145813 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.146005 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.146152 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.136503 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.146603 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.146821 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.146848 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:04.646808694 +0000 UTC m=+21.460425271 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.146935 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.146968 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.147023 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:04.64700294 +0000 UTC m=+21.460619527 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.147058 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.147462 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.147739 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.148000 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.148319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.148639 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.148939 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.149226 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.149962 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.152810 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.156828 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.157257 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.157258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.157900 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.158012 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.158374 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.158803 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.159216 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.159422 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.157983 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.166598 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.166627 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.166644 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.166784 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:04.666752047 +0000 UTC m=+21.480368634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.167097 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.167371 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.182192 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187087 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-ovn\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187130 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-k8s-cni-cncf-io\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187146 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-etc-kubernetes\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187161 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-netns\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-env-overrides\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187192 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcjxt\" (UniqueName: \"kubernetes.io/projected/a4cbf858-76fc-426e-8561-f898927b3bf5-kube-api-access-hcjxt\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187210 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-cnibin\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187224 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-system-cni-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187248 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-slash\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187264 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-multus-certs\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187280 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-proxy-tls\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187296 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-systemd\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187312 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-os-release\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187328 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-cni-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187345 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bcb29cb9-408c-48c2-bdca-2c3dce47b991-cni-binary-copy\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187361 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-etc-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187378 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187398 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-kubelet\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-daemon-config\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187431 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-var-lib-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187448 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-script-lib\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187477 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-netns\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187518 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-cni-bin\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187539 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3b9b7335-d993-46d0-a36a-7821280b7a91-hosts-file\") pod \"node-resolver-vfbbz\" (UID: \"3b9b7335-d993-46d0-a36a-7821280b7a91\") " pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187562 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpqmf\" (UniqueName: \"kubernetes.io/projected/3b9b7335-d993-46d0-a36a-7821280b7a91-kube-api-access-qpqmf\") pod \"node-resolver-vfbbz\" (UID: \"3b9b7335-d993-46d0-a36a-7821280b7a91\") " pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-os-release\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187613 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-netns\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187611 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-conf-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187679 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187692 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-cnibin\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187706 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-node-log\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187726 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-bin\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187737 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-system-cni-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187759 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovn-node-metrics-cert\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187781 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a4cbf858-76fc-426e-8561-f898927b3bf5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187804 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-rootfs\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187823 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187839 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a4cbf858-76fc-426e-8561-f898927b3bf5-cni-binary-copy\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187850 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-ovn\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187897 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-socket-dir-parent\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187920 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-multus-certs\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187927 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-k8s-cni-cncf-io\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187858 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-socket-dir-parent\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187980 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5fkq\" (UniqueName: \"kubernetes.io/projected/bcb29cb9-408c-48c2-bdca-2c3dce47b991-kube-api-access-t5fkq\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188008 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67pdx\" (UniqueName: \"kubernetes.io/projected/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-kube-api-access-67pdx\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188048 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188071 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-netd\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188092 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-config\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188115 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188149 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-systemd-units\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188171 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-log-socket\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188224 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-system-cni-dir\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188247 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-hostroot\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188268 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-kubelet\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188308 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkcm8\" (UniqueName: \"kubernetes.io/projected/f2a63620-dc3c-467f-9cf5-d688f0e233c3-kube-api-access-wkcm8\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188382 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-mcd-auth-proxy-config\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188397 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-env-overrides\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-cnibin\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188426 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-cnibin\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188441 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-cni-multus\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188513 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188529 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188541 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188553 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188566 4696 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188578 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188589 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188601 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188613 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188625 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188637 4696 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188649 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188686 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188702 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188714 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188727 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188740 4696 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188752 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188764 4696 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188776 4696 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188785 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-os-release\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188787 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188823 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188854 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-node-log\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188901 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-cni-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188856 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188927 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-netd\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188898 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188951 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-bin\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188929 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187643 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-conf-dir\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188979 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188993 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189006 4696 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189020 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189032 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189068 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189081 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189093 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189104 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189116 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189128 4696 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189140 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189152 4696 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189169 4696 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189182 4696 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189194 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189207 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189220 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189233 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189246 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189258 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189273 4696 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189285 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189298 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189309 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189321 4696 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189334 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189346 4696 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189358 4696 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189372 4696 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189388 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189400 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189412 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189425 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189437 4696 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189449 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189462 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189474 4696 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189486 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189504 4696 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189516 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189513 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bcb29cb9-408c-48c2-bdca-2c3dce47b991-cni-binary-copy\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189527 4696 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189541 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189553 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189566 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189580 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189592 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187371 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-etc-kubernetes\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189625 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189941 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189959 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189973 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189988 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190000 4696 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190013 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190024 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190037 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190070 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190085 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190096 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190105 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bcb29cb9-408c-48c2-bdca-2c3dce47b991-multus-daemon-config\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190108 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190140 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190152 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190161 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190171 4696 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190180 4696 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190188 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190197 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190207 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190215 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190224 4696 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190233 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190242 4696 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190251 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190260 4696 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190268 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190276 4696 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190285 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190293 4696 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190303 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190311 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190319 4696 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190328 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190336 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190345 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190357 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190366 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190375 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190385 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190396 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190406 4696 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190415 4696 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190424 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190433 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190442 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190452 4696 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190461 4696 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190469 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190479 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190490 4696 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190499 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190509 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190518 4696 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190528 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190537 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190546 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190555 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190564 4696 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190573 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190582 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190592 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190600 4696 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190609 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190617 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190626 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190635 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190644 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190653 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190683 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190694 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190705 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190716 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190730 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190829 4696 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190838 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190846 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190865 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190873 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190892 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190900 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190908 4696 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190916 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190925 4696 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190935 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190944 4696 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190952 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190961 4696 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190970 4696 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190979 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190987 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.190996 4696 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191092 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191104 4696 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191113 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191122 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191130 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191139 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191157 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191166 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191176 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191185 4696 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191194 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191202 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191210 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191219 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191229 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191240 4696 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191252 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.191265 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.194241 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.194263 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.194288 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.194301 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.194313 4696 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188823 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.194379 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-var-lib-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.195346 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-run-netns\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.195423 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-cni-bin\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.195542 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3b9b7335-d993-46d0-a36a-7821280b7a91-hosts-file\") pod \"node-resolver-vfbbz\" (UID: \"3b9b7335-d993-46d0-a36a-7821280b7a91\") " pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.195907 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-os-release\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.188725 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-systemd\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189647 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-kubelet\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.193192 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a4cbf858-76fc-426e-8561-f898927b3bf5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.193237 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-rootfs\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.193267 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.187898 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-slash\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.193957 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-host-var-lib-cni-multus\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.196877 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a4cbf858-76fc-426e-8561-f898927b3bf5-cni-binary-copy\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.196939 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bcb29cb9-408c-48c2-bdca-2c3dce47b991-hostroot\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.197616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.197674 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-systemd-units\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.197728 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-log-socket\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.197760 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4cbf858-76fc-426e-8561-f898927b3bf5-system-cni-dir\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.197972 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.201413 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-kubelet\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.201536 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.189566 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-etc-openvswitch\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.202225 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.202578 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-config\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.202837 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-script-lib\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.203178 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovn-node-metrics-cert\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.203239 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-proxy-tls\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.203366 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-mcd-auth-proxy-config\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.205701 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.208625 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f" exitCode=255 Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.208687 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f"} Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.208726 4696 scope.go:117] "RemoveContainer" containerID="efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.208915 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.209954 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5fkq\" (UniqueName: \"kubernetes.io/projected/bcb29cb9-408c-48c2-bdca-2c3dce47b991-kube-api-access-t5fkq\") pod \"multus-g77vb\" (UID: \"bcb29cb9-408c-48c2-bdca-2c3dce47b991\") " pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.219387 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.219766 4696 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.220486 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcjxt\" (UniqueName: \"kubernetes.io/projected/a4cbf858-76fc-426e-8561-f898927b3bf5-kube-api-access-hcjxt\") pod \"multus-additional-cni-plugins-fb7d4\" (UID: \"a4cbf858-76fc-426e-8561-f898927b3bf5\") " pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.220927 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67pdx\" (UniqueName: \"kubernetes.io/projected/ac9983ff-4a9b-456b-8a9c-7e5857aad5d2-kube-api-access-67pdx\") pod \"machine-config-daemon-2mdzn\" (UID: \"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\") " pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.221082 4696 scope.go:117] "RemoveContainer" containerID="9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.221327 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.222405 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkcm8\" (UniqueName: \"kubernetes.io/projected/f2a63620-dc3c-467f-9cf5-d688f0e233c3-kube-api-access-wkcm8\") pod \"ovnkube-node-l8nb9\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.223957 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpqmf\" (UniqueName: \"kubernetes.io/projected/3b9b7335-d993-46d0-a36a-7821280b7a91-kube-api-access-qpqmf\") pod \"node-resolver-vfbbz\" (UID: \"3b9b7335-d993-46d0-a36a-7821280b7a91\") " pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.224573 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.228720 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.236019 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.262030 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.281142 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.294888 4696 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.294916 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.294927 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.299887 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.308040 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.314050 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:37:04 crc kubenswrapper[4696]: W1125 10:37:04.315227 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-b1e2e07d91f17621c9834627ad236b1e0f38a61537363645be11e8da41d0e759 WatchSource:0}: Error finding container b1e2e07d91f17621c9834627ad236b1e0f38a61537363645be11e8da41d0e759: Status 404 returned error can't find the container with id b1e2e07d91f17621c9834627ad236b1e0f38a61537363645be11e8da41d0e759 Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.320035 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.323889 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vfbbz" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.329032 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.337612 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:04 crc kubenswrapper[4696]: W1125 10:37:04.341510 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-31bed232933dbc7af98706060eac000bbfa8ea2f1989161c2e8236792308efd9 WatchSource:0}: Error finding container 31bed232933dbc7af98706060eac000bbfa8ea2f1989161c2e8236792308efd9: Status 404 returned error can't find the container with id 31bed232933dbc7af98706060eac000bbfa8ea2f1989161c2e8236792308efd9 Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.345637 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.346339 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.351775 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g77vb" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.367339 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.384589 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.402787 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: W1125 10:37:04.408880 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b9b7335_d993_46d0_a36a_7821280b7a91.slice/crio-1fc3dd226ffca601e88438a6f49f89a3675f78f788bd287e6bec5018ea8d7cad WatchSource:0}: Error finding container 1fc3dd226ffca601e88438a6f49f89a3675f78f788bd287e6bec5018ea8d7cad: Status 404 returned error can't find the container with id 1fc3dd226ffca601e88438a6f49f89a3675f78f788bd287e6bec5018ea8d7cad Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.412955 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: W1125 10:37:04.422340 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac9983ff_4a9b_456b_8a9c_7e5857aad5d2.slice/crio-6b93c8ff09c99515ca3b7c9956819794089a6eb39b50af3e4f44482c61e319b0 WatchSource:0}: Error finding container 6b93c8ff09c99515ca3b7c9956819794089a6eb39b50af3e4f44482c61e319b0: Status 404 returned error can't find the container with id 6b93c8ff09c99515ca3b7c9956819794089a6eb39b50af3e4f44482c61e319b0 Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.426341 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: W1125 10:37:04.433525 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcb29cb9_408c_48c2_bdca_2c3dce47b991.slice/crio-ca2654d60d63d6cd4472bd433e645d67c2847b2c0829b9b7a001073d5f86f37f WatchSource:0}: Error finding container ca2654d60d63d6cd4472bd433e645d67c2847b2c0829b9b7a001073d5f86f37f: Status 404 returned error can't find the container with id ca2654d60d63d6cd4472bd433e645d67c2847b2c0829b9b7a001073d5f86f37f Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.460072 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.484494 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.503781 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.525366 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"message\\\":\\\"W1125 10:36:47.259210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 10:36:47.259690 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764067007 cert, and key in /tmp/serving-cert-1039889101/serving-signer.crt, /tmp/serving-cert-1039889101/serving-signer.key\\\\nI1125 10:36:47.460936 1 observer_polling.go:159] Starting file observer\\\\nW1125 10:36:47.464363 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 10:36:47.464798 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:36:47.466711 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1039889101/tls.crt::/tmp/serving-cert-1039889101/tls.key\\\\\\\"\\\\nF1125 10:36:47.725892 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.570297 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.602413 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.620560 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.653658 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"message\\\":\\\"W1125 10:36:47.259210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 10:36:47.259690 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764067007 cert, and key in /tmp/serving-cert-1039889101/serving-signer.crt, /tmp/serving-cert-1039889101/serving-signer.key\\\\nI1125 10:36:47.460936 1 observer_polling.go:159] Starting file observer\\\\nW1125 10:36:47.464363 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 10:36:47.464798 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:36:47.466711 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1039889101/tls.crt::/tmp/serving-cert-1039889101/tls.key\\\\\\\"\\\\nF1125 10:36:47.725892 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.675563 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.692259 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.698242 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.698370 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.698394 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.698412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.698490 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698626 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698641 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698676 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698762 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:05.698734248 +0000 UTC m=+22.512350835 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698831 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:37:05.69881066 +0000 UTC m=+22.512427247 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698870 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698890 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:05.698884552 +0000 UTC m=+22.512501139 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698942 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698953 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698961 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.698980 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:05.698974665 +0000 UTC m=+22.512591252 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.699005 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: E1125 10:37:04.699024 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:05.699017737 +0000 UTC m=+22.512634324 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.713501 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.738788 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.749651 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.760761 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.776454 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.786138 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.894767 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.899050 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.903950 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.905609 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.912153 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.920465 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.928597 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.937349 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.947254 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.956020 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.966766 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.983697 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:04 crc kubenswrapper[4696]: I1125 10:37:04.995318 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.005863 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"message\\\":\\\"W1125 10:36:47.259210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 10:36:47.259690 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764067007 cert, and key in /tmp/serving-cert-1039889101/serving-signer.crt, /tmp/serving-cert-1039889101/serving-signer.key\\\\nI1125 10:36:47.460936 1 observer_polling.go:159] Starting file observer\\\\nW1125 10:36:47.464363 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 10:36:47.464798 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:36:47.466711 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1039889101/tls.crt::/tmp/serving-cert-1039889101/tls.key\\\\\\\"\\\\nF1125 10:36:47.725892 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.014242 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.020416 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.030898 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.041171 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.051144 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.077561 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.116627 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.157240 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc306851ffef25dccae956853419c9dcd7c7402dcaf7ca19168db03bd721545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"message\\\":\\\"W1125 10:36:47.259210 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 10:36:47.259690 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764067007 cert, and key in /tmp/serving-cert-1039889101/serving-signer.crt, /tmp/serving-cert-1039889101/serving-signer.key\\\\nI1125 10:36:47.460936 1 observer_polling.go:159] Starting file observer\\\\nW1125 10:36:47.464363 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 10:36:47.464798 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:36:47.466711 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1039889101/tls.crt::/tmp/serving-cert-1039889101/tls.key\\\\\\\"\\\\nF1125 10:36:47.725892 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.196701 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.212516 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.212743 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b1e2e07d91f17621c9834627ad236b1e0f38a61537363645be11e8da41d0e759"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.213940 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerStarted","Data":"1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.213963 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerStarted","Data":"ca2654d60d63d6cd4472bd433e645d67c2847b2c0829b9b7a001073d5f86f37f"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.217044 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.217079 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"6b93c8ff09c99515ca3b7c9956819794089a6eb39b50af3e4f44482c61e319b0"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.218062 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vfbbz" event={"ID":"3b9b7335-d993-46d0-a36a-7821280b7a91","Type":"ContainerStarted","Data":"1fc3dd226ffca601e88438a6f49f89a3675f78f788bd287e6bec5018ea8d7cad"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.219229 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8" exitCode=0 Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.219282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.219303 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"8a44975297d61e26c42357b632f58b16bebc8697ada169443d0253a9e0c95334"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.225630 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.225732 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f2c6873463d58f57eac147ff2b30edc23cb19cd8f6a59749d40ef0beb65706cc"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.226692 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerStarted","Data":"4d7a540c110c00915a69cd89810203dbbaefb51bf3306f84cbcecf1ec3362d06"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.227544 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"31bed232933dbc7af98706060eac000bbfa8ea2f1989161c2e8236792308efd9"} Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.229228 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.231518 4696 scope.go:117] "RemoveContainer" containerID="9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f" Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.231716 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.243403 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.280653 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.350709 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.365251 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.395708 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.435865 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.477983 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.520045 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.555224 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.597770 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.635716 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.676413 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.709100 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.709191 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.709214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.709241 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.709262 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709365 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709379 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709388 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709426 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:07.709413537 +0000 UTC m=+24.523030124 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709682 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:37:07.709655045 +0000 UTC m=+24.523271632 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709717 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709738 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:07.709732807 +0000 UTC m=+24.523349394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709777 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709786 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709816 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709835 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:07.70982996 +0000 UTC m=+24.523446547 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709871 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: E1125 10:37:05.709888 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:07.709883682 +0000 UTC m=+24.523500269 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.717485 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.767003 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.797150 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.836518 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.874322 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:05 crc kubenswrapper[4696]: I1125 10:37:05.915227 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.042054 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.042091 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.042130 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:06 crc kubenswrapper[4696]: E1125 10:37:06.042196 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:06 crc kubenswrapper[4696]: E1125 10:37:06.042240 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:06 crc kubenswrapper[4696]: E1125 10:37:06.042352 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.045629 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.046244 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.046871 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.047461 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.048045 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.048536 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.049083 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.049620 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.051571 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.052461 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.052997 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.054164 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.054740 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.055274 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.056175 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.057021 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.057991 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.058389 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.058937 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.060035 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.060497 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.061540 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.062106 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.063707 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.064103 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.064733 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.065899 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.066487 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.067908 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.068436 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.069400 4696 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.069503 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.071161 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.072028 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.072913 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.074484 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.075210 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.076142 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.076796 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.077770 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.078256 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.079190 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.079988 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.081015 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.081505 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.082418 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.082983 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.084128 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.084720 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.085811 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.086261 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.087299 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.087891 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.088346 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.234647 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4cbf858-76fc-426e-8561-f898927b3bf5" containerID="77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd" exitCode=0 Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.234735 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerDied","Data":"77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.238366 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.239820 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vfbbz" event={"ID":"3b9b7335-d993-46d0-a36a-7821280b7a91","Type":"ContainerStarted","Data":"edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.243355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.243398 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.243411 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.243421 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.243434 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.243443 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.246849 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90"} Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.257605 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.270831 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.282326 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.297360 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.313622 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.330862 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.341830 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.359953 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.366238 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.374239 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.380673 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.387781 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.391745 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.405936 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.433886 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.460025 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.498135 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.543446 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.578629 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.621776 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.657370 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.702153 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.740859 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.780400 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.819375 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.857497 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.896334 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.937715 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.958807 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.959397 4696 scope.go:117] "RemoveContainer" containerID="9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f" Nov 25 10:37:06 crc kubenswrapper[4696]: E1125 10:37:06.959529 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 10:37:06 crc kubenswrapper[4696]: I1125 10:37:06.976658 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.026339 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.251098 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397"} Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.252883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerStarted","Data":"354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89"} Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.291075 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.306716 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.321963 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.342984 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.359751 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.378609 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.403169 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.419990 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.431468 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.445161 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.459795 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.497374 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.538705 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.578296 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.619185 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.657616 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.696381 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.727852 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728064 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:37:11.728034488 +0000 UTC m=+28.541651075 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.728145 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728311 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728331 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728343 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728386 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:11.728374159 +0000 UTC m=+28.541990746 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728400 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728449 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:11.728438701 +0000 UTC m=+28.542055358 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.728180 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.728651 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.728691 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728734 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728744 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728758 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:11.728751171 +0000 UTC m=+28.542367758 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728763 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728775 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:07 crc kubenswrapper[4696]: E1125 10:37:07.728805 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:11.728794673 +0000 UTC m=+28.542411310 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.738727 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.783634 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.817617 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.859015 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.900058 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.939106 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:07 crc kubenswrapper[4696]: I1125 10:37:07.979306 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:07Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.017111 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.041560 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.041613 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.041568 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:08 crc kubenswrapper[4696]: E1125 10:37:08.041699 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:08 crc kubenswrapper[4696]: E1125 10:37:08.041795 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:08 crc kubenswrapper[4696]: E1125 10:37:08.041873 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.055911 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.106103 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.139583 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.258754 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886"} Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.260214 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4cbf858-76fc-426e-8561-f898927b3bf5" containerID="354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89" exitCode=0 Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.260301 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerDied","Data":"354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89"} Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.272836 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.290023 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.302072 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.317007 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.349257 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.379757 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.421700 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.461133 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.498645 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.542425 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.583893 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.619579 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.659307 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:08 crc kubenswrapper[4696]: I1125 10:37:08.699763 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:08Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.265169 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4cbf858-76fc-426e-8561-f898927b3bf5" containerID="dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6" exitCode=0 Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.265215 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerDied","Data":"dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6"} Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.285398 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.300390 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.311387 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.324167 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.335658 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.347968 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.361361 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.374579 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.384725 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.397282 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.409161 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.429417 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.441651 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.457207 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.473627 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.475380 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.475421 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.475432 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.475515 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.481461 4696 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.481773 4696 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.482681 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.482712 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.482722 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.482738 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.482751 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: E1125 10:37:09.496651 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.500121 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.500159 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.500172 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.500192 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.500203 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: E1125 10:37:09.512731 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.516476 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.516516 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.516528 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.516546 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.516555 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: E1125 10:37:09.528531 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.535960 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.536190 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.536261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.536342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.536404 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: E1125 10:37:09.548649 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.552309 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.552344 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.552354 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.552369 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.552380 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: E1125 10:37:09.563727 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:09Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:09 crc kubenswrapper[4696]: E1125 10:37:09.564008 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.565709 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.565807 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.565961 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.566036 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.566096 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.668851 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.668879 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.668888 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.668901 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.668910 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.771089 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.771124 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.771133 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.771147 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.771158 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.873886 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.873916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.873924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.873937 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.873945 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.976402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.976431 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.976441 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.976459 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:09 crc kubenswrapper[4696]: I1125 10:37:09.976475 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:09Z","lastTransitionTime":"2025-11-25T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.041997 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.042067 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.042124 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:10 crc kubenswrapper[4696]: E1125 10:37:10.042464 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:10 crc kubenswrapper[4696]: E1125 10:37:10.042591 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:10 crc kubenswrapper[4696]: E1125 10:37:10.042755 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.079027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.079064 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.079075 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.079092 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.079103 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.182010 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.182066 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.182077 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.182095 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.182104 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.271772 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4cbf858-76fc-426e-8561-f898927b3bf5" containerID="bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d" exitCode=0 Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.271992 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerDied","Data":"bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.285086 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.285163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.285175 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.285200 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.285215 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.291968 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.312205 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.340416 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.355768 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.367796 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.383542 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.388213 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.388254 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.388264 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.388286 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.388300 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.397497 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.413603 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.435889 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.447160 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.463950 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.487945 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.491579 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.491646 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.491695 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.491729 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.491749 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.504511 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.522909 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.594179 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.594217 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.594227 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.594241 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.594251 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.696633 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.696698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.696712 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.696733 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.696748 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.799377 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.799425 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.799439 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.799459 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.799471 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.902005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.902058 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.902073 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.902097 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:10 crc kubenswrapper[4696]: I1125 10:37:10.902110 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:10Z","lastTransitionTime":"2025-11-25T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.005641 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.005693 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.005706 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.005722 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.005732 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.107978 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.108010 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.108020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.108035 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.108046 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.210528 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.210568 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.210578 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.210595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.210605 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.279163 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerStarted","Data":"9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.283522 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.284379 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.284450 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.309950 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.312448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.312489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.312511 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.312531 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.312546 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.327972 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.339290 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.341099 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.348866 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.367696 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.383716 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.403235 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.413896 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.415134 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.415159 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.415169 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.415181 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.415190 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.425997 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.438715 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.455590 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.467628 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.477249 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.488684 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.501585 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.514908 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.517838 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.517885 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.517896 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.517914 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.517925 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.534513 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.545940 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.561945 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.578007 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.595479 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.609474 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.620891 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.620943 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.620970 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.620990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.621001 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.624125 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.635618 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.646381 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.658061 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.676169 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.687254 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.700427 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.723390 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.723455 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.723467 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.723483 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.723493 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.765821 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.765953 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766036 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:37:19.766006254 +0000 UTC m=+36.579622841 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766081 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.766157 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.766209 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766257 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:19.766220711 +0000 UTC m=+36.579837298 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766289 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.766323 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766347 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766367 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766379 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766349 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:19.766326694 +0000 UTC m=+36.579943331 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766423 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:19.766413337 +0000 UTC m=+36.580029924 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766453 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766475 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766487 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:11 crc kubenswrapper[4696]: E1125 10:37:11.766530 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:19.76651596 +0000 UTC m=+36.580132547 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.770365 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gk2rz"] Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.770781 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.772839 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.772995 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.773260 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.773305 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.785348 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.801153 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.814457 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.826490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.826525 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.826535 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.826550 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.826559 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.837933 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.850320 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.863003 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.867419 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-serviceca\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.867465 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-host\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.867495 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pdxg\" (UniqueName: \"kubernetes.io/projected/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-kube-api-access-5pdxg\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.874995 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.887444 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.899440 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.917162 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.929267 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.929313 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.929325 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.929341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.929780 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:11Z","lastTransitionTime":"2025-11-25T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.929865 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.941340 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.949375 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.961311 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.968791 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pdxg\" (UniqueName: \"kubernetes.io/projected/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-kube-api-access-5pdxg\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.968852 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-serviceca\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.968870 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-host\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.968909 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-host\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.969871 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-serviceca\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.972990 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:11 crc kubenswrapper[4696]: I1125 10:37:11.984692 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pdxg\" (UniqueName: \"kubernetes.io/projected/173afc30-0ef2-4c6d-b6fa-48c7905d16dc-kube-api-access-5pdxg\") pod \"node-ca-gk2rz\" (UID: \"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\") " pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.031738 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.031772 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.031782 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.031796 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.031804 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.042070 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:12 crc kubenswrapper[4696]: E1125 10:37:12.042163 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.042232 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.042269 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:12 crc kubenswrapper[4696]: E1125 10:37:12.042342 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:12 crc kubenswrapper[4696]: E1125 10:37:12.042414 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.084101 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gk2rz" Nov 25 10:37:12 crc kubenswrapper[4696]: W1125 10:37:12.098583 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod173afc30_0ef2_4c6d_b6fa_48c7905d16dc.slice/crio-f2ca7afa52da471fe5dc3f952d00c23917d752ac53540f77e93394707f6ce5b2 WatchSource:0}: Error finding container f2ca7afa52da471fe5dc3f952d00c23917d752ac53540f77e93394707f6ce5b2: Status 404 returned error can't find the container with id f2ca7afa52da471fe5dc3f952d00c23917d752ac53540f77e93394707f6ce5b2 Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.134736 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.135006 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.135096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.135180 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.135281 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.237943 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.237980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.237991 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.238007 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.238016 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.286915 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gk2rz" event={"ID":"173afc30-0ef2-4c6d-b6fa-48c7905d16dc","Type":"ContainerStarted","Data":"f2ca7afa52da471fe5dc3f952d00c23917d752ac53540f77e93394707f6ce5b2"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.290136 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4cbf858-76fc-426e-8561-f898927b3bf5" containerID="9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e" exitCode=0 Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.290174 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerDied","Data":"9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.290499 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.318060 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.338330 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.340058 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.340090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.340097 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.340112 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.340121 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.352621 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.363871 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.376063 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.393222 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.406596 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.422226 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.436865 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.442542 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.442571 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.442580 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.442595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.442603 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.449336 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.461141 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.474625 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.484871 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.499111 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.514776 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:12Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.544444 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.544480 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.544491 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.544506 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.544518 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.647011 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.647053 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.647062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.647076 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.647089 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.749772 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.749822 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.749833 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.749851 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.749863 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.852061 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.852112 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.852123 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.852140 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.852155 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.953958 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.954293 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.954306 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.954521 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:12 crc kubenswrapper[4696]: I1125 10:37:12.954533 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:12Z","lastTransitionTime":"2025-11-25T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.056449 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.056486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.056497 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.056513 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.056524 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.159122 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.159158 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.159168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.159187 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.159215 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.261014 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.261045 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.261053 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.261066 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.261075 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.295642 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4cbf858-76fc-426e-8561-f898927b3bf5" containerID="8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60" exitCode=0 Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.295703 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerDied","Data":"8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.297996 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gk2rz" event={"ID":"173afc30-0ef2-4c6d-b6fa-48c7905d16dc","Type":"ContainerStarted","Data":"4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.298093 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.311082 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.325392 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.337210 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.357496 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.363355 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.363380 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.363389 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.363401 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.363409 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.369482 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.384210 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.398132 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.413991 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.427899 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.452150 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.464861 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.464882 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.464889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.464901 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.464909 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.479721 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.507853 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.519473 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.535502 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.546495 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.564413 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.566552 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.566586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.566599 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.566636 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.566649 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.578336 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.595632 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.605977 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.618796 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.631444 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.643962 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.655508 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.667912 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.668551 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.668595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.668606 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.668621 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.668630 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.687002 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.699124 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.711995 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.726705 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.739574 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.754057 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:13Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.770619 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.770659 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.770687 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.770702 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.770716 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.873837 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.873883 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.873892 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.873907 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.873916 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.918202 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.976308 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.976349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.976357 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.976371 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:13 crc kubenswrapper[4696]: I1125 10:37:13.976380 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:13Z","lastTransitionTime":"2025-11-25T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.041328 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.041349 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.041397 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:14 crc kubenswrapper[4696]: E1125 10:37:14.041827 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:14 crc kubenswrapper[4696]: E1125 10:37:14.042118 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:14 crc kubenswrapper[4696]: E1125 10:37:14.042240 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.062092 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.074460 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.078363 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.078553 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.078625 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.078722 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.078800 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.083712 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.094255 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.103138 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.121091 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.132116 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.147099 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.161560 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.173150 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.180438 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.180466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.180479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.180495 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.180506 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.184409 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.200857 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.212547 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.224655 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.237986 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.282617 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.282658 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.282698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.282728 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.282751 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.303270 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/0.log" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.307896 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75" exitCode=1 Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.308060 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.309259 4696 scope.go:117] "RemoveContainer" containerID="397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.316557 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" event={"ID":"a4cbf858-76fc-426e-8561-f898927b3bf5","Type":"ContainerStarted","Data":"49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.326989 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.349843 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.370494 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.384999 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.385033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.385047 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.385066 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.385078 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.385880 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.401819 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.419391 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 10:37:13.582216 5818 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 10:37:13.582233 5818 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 10:37:13.582265 5818 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 10:37:13.582287 5818 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 10:37:13.582299 5818 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 10:37:13.582325 5818 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 10:37:13.582334 5818 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 10:37:13.582355 5818 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:37:13.582384 5818 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 10:37:13.582396 5818 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 10:37:13.582359 5818 factory.go:656] Stopping watch factory\\\\nI1125 10:37:13.582418 5818 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:37:13.582368 5818 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 10:37:13.582456 5818 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 10:37:13.582468 5818 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.433876 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.448372 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.460385 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.481843 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.487295 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.487321 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.487329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.487343 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.487352 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.518736 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.566627 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.590033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.590063 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.590071 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.590086 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.590097 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.598605 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.643870 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.679189 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.691727 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.691755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.691765 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.691778 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.691787 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.717784 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.756034 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.794208 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.794252 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.794265 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.794286 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.794298 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.802868 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.837922 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.885687 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.896792 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.896839 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.896849 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.896864 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.896873 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.919965 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.960441 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.995806 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.999393 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.999419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.999428 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.999458 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:14 crc kubenswrapper[4696]: I1125 10:37:14.999467 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:14Z","lastTransitionTime":"2025-11-25T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.045340 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 10:37:13.582216 5818 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 10:37:13.582233 5818 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 10:37:13.582265 5818 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 10:37:13.582287 5818 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 10:37:13.582299 5818 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 10:37:13.582325 5818 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 10:37:13.582334 5818 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 10:37:13.582355 5818 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:37:13.582384 5818 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 10:37:13.582396 5818 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 10:37:13.582359 5818 factory.go:656] Stopping watch factory\\\\nI1125 10:37:13.582418 5818 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:37:13.582368 5818 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 10:37:13.582456 5818 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 10:37:13.582468 5818 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.077957 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.101482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.101526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.101540 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.101556 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.101567 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.117909 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.158962 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.197839 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.203595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.203639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.203650 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.203690 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.203704 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.236992 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.277628 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.306081 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.306115 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.306128 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.306150 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.306162 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.324112 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/1.log" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.324872 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/0.log" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.327444 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54" exitCode=1 Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.327491 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.327555 4696 scope.go:117] "RemoveContainer" containerID="397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.328739 4696 scope.go:117] "RemoveContainer" containerID="e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54" Nov 25 10:37:15 crc kubenswrapper[4696]: E1125 10:37:15.328971 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.344446 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.362449 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.399509 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.408399 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.408436 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.408447 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.408462 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.408473 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.443843 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.475951 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.510972 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.511000 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.511008 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.511021 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.511029 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.520294 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.560171 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.604927 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://397fca4bb62bafc9d05bd6955959943e3caf9b3eb63043cf8784caf4f7c61b75\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 10:37:13.582216 5818 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 10:37:13.582233 5818 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 10:37:13.582265 5818 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 10:37:13.582287 5818 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 10:37:13.582299 5818 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 10:37:13.582325 5818 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 10:37:13.582334 5818 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 10:37:13.582355 5818 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:37:13.582384 5818 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 10:37:13.582396 5818 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 10:37:13.582359 5818 factory.go:656] Stopping watch factory\\\\nI1125 10:37:13.582418 5818 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:37:13.582368 5818 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 10:37:13.582456 5818 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 10:37:13.582468 5818 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.613579 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.613713 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.613725 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.613742 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.613752 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.642786 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.679948 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.715607 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.715693 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.715707 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.715724 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.716036 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.719683 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.760194 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.800072 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.817866 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.817900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.817910 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.817923 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.817932 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.835596 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.876522 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.920618 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.920927 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.921024 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.921157 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:15 crc kubenswrapper[4696]: I1125 10:37:15.921255 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:15Z","lastTransitionTime":"2025-11-25T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.024046 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.024256 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.024372 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.024464 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.024545 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.041890 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.041965 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:16 crc kubenswrapper[4696]: E1125 10:37:16.042026 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.042106 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:16 crc kubenswrapper[4696]: E1125 10:37:16.042247 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:16 crc kubenswrapper[4696]: E1125 10:37:16.042365 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.127958 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.128005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.128017 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.128036 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.128048 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.230349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.230389 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.230402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.230421 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.230433 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.332276 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.332307 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.332327 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.332343 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.332354 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.333353 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/1.log" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.336622 4696 scope.go:117] "RemoveContainer" containerID="e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54" Nov 25 10:37:16 crc kubenswrapper[4696]: E1125 10:37:16.336830 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.359655 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.380891 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.398581 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.413422 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.422422 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.434485 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.434662 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.434749 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.434816 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.434870 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.435292 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.451075 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.473766 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.494659 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.508880 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.521913 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.528035 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k"] Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.528513 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.530179 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.531637 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.536976 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.537562 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.537697 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.537787 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.537880 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.538223 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.553889 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.564641 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.574774 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.589926 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.602086 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.612157 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f3ebd497-beb3-40a3-9c6a-8ad74b514311-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.612216 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f3ebd497-beb3-40a3-9c6a-8ad74b514311-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.612242 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f3ebd497-beb3-40a3-9c6a-8ad74b514311-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.612264 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckngr\" (UniqueName: \"kubernetes.io/projected/f3ebd497-beb3-40a3-9c6a-8ad74b514311-kube-api-access-ckngr\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.640096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.640123 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.640131 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.640144 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.640152 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.644469 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.680562 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.713544 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f3ebd497-beb3-40a3-9c6a-8ad74b514311-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.713631 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f3ebd497-beb3-40a3-9c6a-8ad74b514311-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.713657 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckngr\" (UniqueName: \"kubernetes.io/projected/f3ebd497-beb3-40a3-9c6a-8ad74b514311-kube-api-access-ckngr\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.713712 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f3ebd497-beb3-40a3-9c6a-8ad74b514311-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.714484 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f3ebd497-beb3-40a3-9c6a-8ad74b514311-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.714714 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f3ebd497-beb3-40a3-9c6a-8ad74b514311-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.721925 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f3ebd497-beb3-40a3-9c6a-8ad74b514311-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.722447 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.742638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.742996 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.743061 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.743162 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.743232 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.745939 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckngr\" (UniqueName: \"kubernetes.io/projected/f3ebd497-beb3-40a3-9c6a-8ad74b514311-kube-api-access-ckngr\") pod \"ovnkube-control-plane-749d76644c-q244k\" (UID: \"f3ebd497-beb3-40a3-9c6a-8ad74b514311\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.780188 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.820479 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.840522 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.846574 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.846616 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.846635 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.846734 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.846754 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: W1125 10:37:16.854917 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3ebd497_beb3_40a3_9c6a_8ad74b514311.slice/crio-1a9f8baa2e6271afadf2e49bf000eba94f60e22e7cbdf1274874434715b73c45 WatchSource:0}: Error finding container 1a9f8baa2e6271afadf2e49bf000eba94f60e22e7cbdf1274874434715b73c45: Status 404 returned error can't find the container with id 1a9f8baa2e6271afadf2e49bf000eba94f60e22e7cbdf1274874434715b73c45 Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.858352 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.900244 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.941581 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.949691 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.949754 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.949769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.949797 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.949813 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:16Z","lastTransitionTime":"2025-11-25T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:16 crc kubenswrapper[4696]: I1125 10:37:16.978546 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.023123 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.052013 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.052066 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.052076 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.052123 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.052133 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.064578 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.096632 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.142864 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.153966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.154003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.154013 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.154027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.154037 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.178078 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.248036 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-w7l9g"] Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.248519 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: E1125 10:37:17.248585 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.255855 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.255894 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.255904 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.255919 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.255930 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.268038 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.303561 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.320109 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.320156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n27ht\" (UniqueName: \"kubernetes.io/projected/19727774-75c0-47f9-9157-75b1116d7daa-kube-api-access-n27ht\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.328476 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.345159 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.345537 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" event={"ID":"f3ebd497-beb3-40a3-9c6a-8ad74b514311","Type":"ContainerStarted","Data":"2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.345615 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" event={"ID":"f3ebd497-beb3-40a3-9c6a-8ad74b514311","Type":"ContainerStarted","Data":"1a9f8baa2e6271afadf2e49bf000eba94f60e22e7cbdf1274874434715b73c45"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.358541 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.358579 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.358589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.358604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.358614 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.379558 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.421134 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.421170 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n27ht\" (UniqueName: \"kubernetes.io/projected/19727774-75c0-47f9-9157-75b1116d7daa-kube-api-access-n27ht\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: E1125 10:37:17.421320 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:17 crc kubenswrapper[4696]: E1125 10:37:17.421410 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:37:17.921391734 +0000 UTC m=+34.735008321 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.427429 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.444656 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n27ht\" (UniqueName: \"kubernetes.io/projected/19727774-75c0-47f9-9157-75b1116d7daa-kube-api-access-n27ht\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.460867 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.460904 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.460916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.460932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.460944 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.478296 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.519726 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.555406 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.563087 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.563124 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.563137 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.563153 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.563164 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.597384 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.637114 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.665883 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.665924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.665936 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.665953 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.665965 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.682063 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.719257 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.758609 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.767505 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.767540 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.767550 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.767564 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.767575 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.798461 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.837476 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.869418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.869448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.869457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.869472 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.869481 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.877171 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:17Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.926025 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:17 crc kubenswrapper[4696]: E1125 10:37:17.926195 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:17 crc kubenswrapper[4696]: E1125 10:37:17.926250 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:37:18.926233763 +0000 UTC m=+35.739850350 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.971940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.971982 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.971991 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.972004 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:17 crc kubenswrapper[4696]: I1125 10:37:17.972013 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:17Z","lastTransitionTime":"2025-11-25T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.042325 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.042377 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.042336 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:18 crc kubenswrapper[4696]: E1125 10:37:18.042467 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:18 crc kubenswrapper[4696]: E1125 10:37:18.042526 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:18 crc kubenswrapper[4696]: E1125 10:37:18.042623 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.074587 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.074643 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.074652 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.074686 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.074696 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.177354 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.177388 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.177397 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.177413 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.177421 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.279254 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.279291 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.279298 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.279312 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.279321 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.353143 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" event={"ID":"f3ebd497-beb3-40a3-9c6a-8ad74b514311","Type":"ContainerStarted","Data":"52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.365969 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.379131 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.381873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.381905 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.381918 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.381934 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.381949 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.393771 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.403763 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.424635 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.440434 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.453318 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.466178 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.477341 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.484098 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.484137 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.484147 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.484164 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.484173 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.494795 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.505829 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.519204 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.530638 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.543182 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.551988 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.560770 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.570483 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:18Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.585812 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.586026 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.586140 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.586328 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.586438 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.688852 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.689205 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.689342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.689475 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.689602 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.792087 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.792140 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.792160 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.792181 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.792199 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.894891 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.894946 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.894962 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.894981 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.894997 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.937252 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:18 crc kubenswrapper[4696]: E1125 10:37:18.937452 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:18 crc kubenswrapper[4696]: E1125 10:37:18.937700 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:37:20.937634736 +0000 UTC m=+37.751251393 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.997344 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.997378 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.997387 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.997401 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:18 crc kubenswrapper[4696]: I1125 10:37:18.997410 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:18Z","lastTransitionTime":"2025-11-25T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.041720 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.041880 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.100565 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.100603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.100612 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.100627 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.100639 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.203751 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.203798 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.203810 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.203830 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.203841 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.306339 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.306418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.306429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.306449 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.306461 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.408819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.409863 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.409992 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.410111 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.410222 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.513677 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.513709 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.513717 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.513729 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.513737 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.616862 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.616897 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.616905 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.616919 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.616928 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.719898 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.719983 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.719994 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.720008 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.720018 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.751758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.751812 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.751824 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.751843 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.751853 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.767306 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:19Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.771181 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.771221 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.771230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.771258 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.771279 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.782899 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:19Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.786347 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.786408 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.786426 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.786452 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.786467 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.799733 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:19Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.808052 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.808090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.808100 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.808116 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.808127 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.821896 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:19Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.824926 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.824964 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.824973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.824989 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.824998 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.836255 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:19Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.836419 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.837887 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.837921 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.837930 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.837943 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.837952 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.846294 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.846395 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846466 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:37:35.84644456 +0000 UTC m=+52.660061147 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846496 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.846530 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846545 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:35.846532423 +0000 UTC m=+52.660149010 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.846573 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.846598 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846718 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846733 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846743 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846748 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846719 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846784 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846792 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846774 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:35.846766021 +0000 UTC m=+52.660382608 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846824 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:35.846814652 +0000 UTC m=+52.660431239 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:19 crc kubenswrapper[4696]: E1125 10:37:19.846835 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:37:35.846830533 +0000 UTC m=+52.660447120 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.940466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.940544 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.940561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.940581 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:19 crc kubenswrapper[4696]: I1125 10:37:19.940594 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:19Z","lastTransitionTime":"2025-11-25T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.041458 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:20 crc kubenswrapper[4696]: E1125 10:37:20.041591 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.041597 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.041426 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:20 crc kubenswrapper[4696]: E1125 10:37:20.042322 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:20 crc kubenswrapper[4696]: E1125 10:37:20.042532 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.042850 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.042874 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.042882 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.042892 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.042901 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.043113 4696 scope.go:117] "RemoveContainer" containerID="9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.145549 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.145610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.145619 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.145635 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.145646 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.248230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.248265 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.248274 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.248290 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.248299 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.351568 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.351621 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.351632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.351651 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.351683 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.362545 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.364423 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.364724 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.385107 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.401080 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.417448 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.430435 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.442342 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.457139 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.457194 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.457206 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.457227 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.457237 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.458019 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.480117 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.492265 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.505057 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.518912 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.538264 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.553755 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.559878 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.559928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.559940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.559954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.559967 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.572076 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.591921 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.608446 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.625276 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.636540 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.662408 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.662435 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.662443 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.662458 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.662470 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.765331 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.765400 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.765425 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.765453 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.765474 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.868221 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.868284 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.868297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.868336 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.868349 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.958860 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:20 crc kubenswrapper[4696]: E1125 10:37:20.959082 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:20 crc kubenswrapper[4696]: E1125 10:37:20.959356 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:37:24.959330265 +0000 UTC m=+41.772946872 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.970931 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.970993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.971014 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.971044 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:20 crc kubenswrapper[4696]: I1125 10:37:20.971068 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:20Z","lastTransitionTime":"2025-11-25T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.042183 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:21 crc kubenswrapper[4696]: E1125 10:37:21.042391 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.073989 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.074064 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.074088 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.074117 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.074140 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.178048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.178342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.178420 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.178490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.178552 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.280395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.280432 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.280442 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.280457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.280469 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.382412 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.382469 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.382481 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.382496 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.382507 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.485364 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.485423 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.485445 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.485471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.485487 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.587507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.587841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.587973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.588065 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.588153 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.691876 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.692203 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.692392 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.692540 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.692715 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.796083 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.796124 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.796135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.796152 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.796162 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.899582 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.899635 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.899649 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.899690 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:21 crc kubenswrapper[4696]: I1125 10:37:21.899705 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:21Z","lastTransitionTime":"2025-11-25T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.002564 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.002907 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.003086 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.003226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.003355 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.041723 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.041743 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:22 crc kubenswrapper[4696]: E1125 10:37:22.041892 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:22 crc kubenswrapper[4696]: E1125 10:37:22.041994 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.041739 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:22 crc kubenswrapper[4696]: E1125 10:37:22.042538 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.106828 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.106902 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.106916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.106940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.106953 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.209573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.209612 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.209622 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.209638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.209649 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.312261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.312299 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.312310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.312326 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.312336 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.415824 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.415873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.415886 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.415903 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.415915 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.518923 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.518974 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.518985 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.519005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.519319 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.621587 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.621632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.621658 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.621704 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.621719 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.724018 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.724062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.724074 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.724092 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.724104 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.826948 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.827005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.827017 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.827037 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.827049 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.929542 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.929594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.929603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.929616 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:22 crc kubenswrapper[4696]: I1125 10:37:22.929628 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:22Z","lastTransitionTime":"2025-11-25T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.032020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.032064 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.032078 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.032103 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.032114 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.041339 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:23 crc kubenswrapper[4696]: E1125 10:37:23.041467 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.134391 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.134433 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.134444 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.134462 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.134474 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.236867 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.236909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.236919 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.236935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.236948 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.340230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.340297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.340313 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.340338 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.340353 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.444557 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.444613 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.444623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.444638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.444650 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.547121 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.547160 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.547170 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.547184 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.547194 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.650411 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.650477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.650499 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.650526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.650544 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.753521 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.753577 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.753589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.753607 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.753619 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.855795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.855854 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.855863 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.855877 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.855886 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.958343 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.958395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.958406 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.958426 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:23 crc kubenswrapper[4696]: I1125 10:37:23.958438 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:23Z","lastTransitionTime":"2025-11-25T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.041205 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.041305 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:24 crc kubenswrapper[4696]: E1125 10:37:24.041445 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.041607 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:24 crc kubenswrapper[4696]: E1125 10:37:24.041745 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:24 crc kubenswrapper[4696]: E1125 10:37:24.041799 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.060401 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.060430 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.060438 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.060451 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.060460 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.068781 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.084341 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.101243 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.112927 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.125701 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.141484 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.157018 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.162466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.162700 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.162781 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.162892 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.162968 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.168119 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.181722 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.193317 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.203309 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.212606 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.223787 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.233259 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.243902 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.254389 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.264716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.264748 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.264758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.264771 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.264780 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.266553 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.366637 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.366706 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.366721 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.366741 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.366758 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.469165 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.469209 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.469220 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.469237 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.469249 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.571715 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.571746 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.571755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.571770 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.571781 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.674691 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.674971 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.675051 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.675130 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.675240 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.777621 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.777652 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.777674 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.777687 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.777709 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.880255 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.880289 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.880299 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.880314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.880325 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.982323 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.982365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.982375 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.982391 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:24 crc kubenswrapper[4696]: I1125 10:37:24.982401 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:24Z","lastTransitionTime":"2025-11-25T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.005243 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:25 crc kubenswrapper[4696]: E1125 10:37:25.005427 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:25 crc kubenswrapper[4696]: E1125 10:37:25.005543 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:37:33.005519703 +0000 UTC m=+49.819136380 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.042148 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:25 crc kubenswrapper[4696]: E1125 10:37:25.042282 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.085048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.085079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.085089 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.085104 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.085114 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.187312 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.187360 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.187372 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.187392 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.187405 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.290323 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.290709 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.290909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.291116 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.291336 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.394110 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.394191 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.394215 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.394246 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.394272 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.497498 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.497543 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.497555 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.497573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.497586 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.600949 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.601009 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.601030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.601057 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.601079 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.704130 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.704183 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.704202 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.704225 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.704242 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.807210 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.807260 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.807271 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.807289 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.807301 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.909235 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.909267 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.909275 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.909288 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:25 crc kubenswrapper[4696]: I1125 10:37:25.909297 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:25Z","lastTransitionTime":"2025-11-25T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.011492 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.011526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.011534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.011547 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.011555 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.042202 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.042290 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:26 crc kubenswrapper[4696]: E1125 10:37:26.042341 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.042358 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:26 crc kubenswrapper[4696]: E1125 10:37:26.042438 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:26 crc kubenswrapper[4696]: E1125 10:37:26.042596 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.113980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.114021 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.114033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.114049 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.114059 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.216550 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.216595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.216608 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.216627 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.216642 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.318877 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.318924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.318940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.318961 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.318977 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.420870 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.420927 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.420943 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.420966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.420978 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.523307 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.523354 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.523372 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.523394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.523410 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.625730 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.625755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.625764 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.625778 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.625786 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.728203 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.728246 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.728260 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.728280 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.728294 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.830710 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.830768 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.830821 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.830845 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.830856 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.933565 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.933606 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.933618 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.933632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:26 crc kubenswrapper[4696]: I1125 10:37:26.933641 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:26Z","lastTransitionTime":"2025-11-25T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.036399 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.036441 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.036452 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.036467 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.036477 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.041690 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:27 crc kubenswrapper[4696]: E1125 10:37:27.041795 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.139281 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.139333 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.139383 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.139402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.139414 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.241281 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.241416 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.241437 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.241453 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.241462 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.344259 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.344314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.344324 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.344342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.344354 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.446395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.446435 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.446444 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.446459 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.446474 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.548990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.549028 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.549037 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.549051 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.549060 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.651541 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.651580 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.651591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.651611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.651623 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.754397 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.754445 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.754458 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.754476 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.754487 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.856842 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.856879 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.856888 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.856902 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.856911 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.959561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.959594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.959602 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.959615 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:27 crc kubenswrapper[4696]: I1125 10:37:27.959624 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:27Z","lastTransitionTime":"2025-11-25T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.041274 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.041405 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:28 crc kubenswrapper[4696]: E1125 10:37:28.041556 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.041607 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:28 crc kubenswrapper[4696]: E1125 10:37:28.041742 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:28 crc kubenswrapper[4696]: E1125 10:37:28.041793 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.061491 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.061536 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.061549 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.061567 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.061579 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.164103 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.164147 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.164159 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.164176 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.164188 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.266293 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.266336 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.266348 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.266363 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.266373 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.369048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.369105 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.369114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.369131 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.369143 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.471886 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.471922 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.471932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.471966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.471977 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.574369 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.574441 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.574458 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.574482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.574504 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.676395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.676448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.676457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.676470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.676478 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.779409 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.779450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.779460 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.779479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.779495 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.881803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.881838 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.881849 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.881866 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.881880 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.984525 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.984817 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.984827 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.984842 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:28 crc kubenswrapper[4696]: I1125 10:37:28.984853 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:28Z","lastTransitionTime":"2025-11-25T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.041508 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:29 crc kubenswrapper[4696]: E1125 10:37:29.041628 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.087533 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.087590 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.087611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.087639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.087696 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.189844 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.189872 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.189880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.189893 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.189901 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.292819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.293114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.293178 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.293261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.293361 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.395897 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.395952 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.395974 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.396021 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.396042 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.498551 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.498601 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.498614 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.498632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.498644 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.600879 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.600939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.600954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.600972 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.600986 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.703639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.703754 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.703771 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.703795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.703814 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.806708 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.806751 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.806762 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.806780 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.806792 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.909155 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.909191 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.909201 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.909281 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:29 crc kubenswrapper[4696]: I1125 10:37:29.909293 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:29Z","lastTransitionTime":"2025-11-25T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.012365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.012446 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.012478 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.012520 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.012539 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.041208 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.041258 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.041255 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.041348 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.041504 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.041585 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.115138 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.115405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.115462 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.115486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.115506 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.119074 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.119130 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.119142 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.119160 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.119172 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.134306 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.138218 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.138604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.138834 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.138987 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.139074 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.152554 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.156236 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.156319 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.156333 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.156349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.156361 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.169747 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.174063 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.174118 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.174125 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.174141 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.174166 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.185919 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.189825 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.189862 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.189873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.189890 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.189904 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.202156 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:30 crc kubenswrapper[4696]: E1125 10:37:30.202374 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.218904 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.218954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.218963 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.218977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.218986 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.321478 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.321523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.321533 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.321549 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.321562 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.424129 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.424180 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.424194 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.424210 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.424222 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.527187 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.527247 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.527263 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.527284 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.527296 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.630550 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.630631 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.630652 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.630723 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.630748 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.733788 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.733866 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.733888 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.733921 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.733943 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.836287 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.836453 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.836478 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.836499 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.836515 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.939522 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.940407 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.940516 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.940548 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:30 crc kubenswrapper[4696]: I1125 10:37:30.940562 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:30Z","lastTransitionTime":"2025-11-25T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.041459 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:31 crc kubenswrapper[4696]: E1125 10:37:31.041778 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.042871 4696 scope.go:117] "RemoveContainer" containerID="e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.043696 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.043741 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.043761 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.043782 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.043803 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.146361 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.146390 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.146399 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.146411 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.146420 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.248965 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.248992 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.249000 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.249015 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.249024 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.351766 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.351803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.351814 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.351829 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.351840 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.395748 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/1.log" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.398484 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.399536 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.435336 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.454035 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.454252 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.454331 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.454138 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.454404 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.454545 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.467138 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.479978 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.491221 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.502918 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.516205 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.529804 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.543385 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.554027 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.556461 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.556507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.556518 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.556535 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.556546 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.566741 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.579164 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.594387 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.612866 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.626944 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.643485 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.652946 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.658278 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.658310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.658319 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.658334 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.658345 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.760141 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.760624 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.760708 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.760772 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.760836 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.863261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.863574 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.863644 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.863786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.863855 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.966419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.966647 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.966767 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.966846 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:31 crc kubenswrapper[4696]: I1125 10:37:31.966946 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:31Z","lastTransitionTime":"2025-11-25T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.041792 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:32 crc kubenswrapper[4696]: E1125 10:37:32.041930 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.041792 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:32 crc kubenswrapper[4696]: E1125 10:37:32.042043 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.041806 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:32 crc kubenswrapper[4696]: E1125 10:37:32.042138 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.069590 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.069655 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.069688 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.069712 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.069731 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.171834 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.172106 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.172171 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.172293 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.172383 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.275230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.275292 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.275309 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.275325 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.275334 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.378255 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.378285 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.378293 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.378306 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.378315 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.403706 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/2.log" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.404498 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/1.log" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.407128 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd" exitCode=1 Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.407297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.407437 4696 scope.go:117] "RemoveContainer" containerID="e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.408407 4696 scope.go:117] "RemoveContainer" containerID="c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd" Nov 25 10:37:32 crc kubenswrapper[4696]: E1125 10:37:32.409710 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.432684 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.442611 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.453812 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.464116 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.476504 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.480327 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.480362 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.480375 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.480394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.480409 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.500123 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.518115 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.532907 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.541435 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.551818 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.568484 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.581898 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.582297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.582365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.582378 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.582396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.582409 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.593576 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.610392 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.629365 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c028d52526ca244ea241e9c59c5f2599cb6a3c8f72d1ab1b7150fcc9285e54\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:15Z\\\",\\\"message\\\":\\\"network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:15Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:15.038926 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-network-diagnostics Admin Network Policy controller: took 3.96µs\\\\nI1125 10:37:15.038932 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-ovirt-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038935 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-ovirt-infra Admin Network Policy controller: took 3.56µs\\\\nI1125 10:37:15.038939 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-cluster-version in Admin Network Policy controller\\\\nI1125 10:37:15.038942 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-cluster-version Admin Network Policy controller: took 3.561µs\\\\nI1125 10:37:15.038947 6042 admin_network_policy_namespace.go:53] Processing sync for Namespace openshift-infra in Admin Network Policy controller\\\\nI1125 10:37:15.038950 6042 admin_network_policy_namespace.go:56] Finished syncing Namespace openshift-infra Admin Network Policy controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.643333 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.656488 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.684382 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.684629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.684742 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.684817 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.684875 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.787650 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.787730 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.787740 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.787754 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.787766 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.890092 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.890139 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.890152 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.890167 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.890179 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.992776 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.992825 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.992840 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.992860 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:32 crc kubenswrapper[4696]: I1125 10:37:32.992871 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:32Z","lastTransitionTime":"2025-11-25T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.041890 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:33 crc kubenswrapper[4696]: E1125 10:37:33.042033 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.084208 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:33 crc kubenswrapper[4696]: E1125 10:37:33.084415 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:33 crc kubenswrapper[4696]: E1125 10:37:33.084509 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:37:49.084486304 +0000 UTC m=+65.898102921 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.095224 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.095283 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.095304 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.095335 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.095358 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.197763 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.197810 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.197824 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.197841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.197853 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.300420 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.300461 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.300473 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.300491 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.300505 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.403452 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.403497 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.403508 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.403526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.403538 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.413254 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/2.log" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.418239 4696 scope.go:117] "RemoveContainer" containerID="c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd" Nov 25 10:37:33 crc kubenswrapper[4696]: E1125 10:37:33.418593 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.442762 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.456599 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.473118 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.486582 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.499648 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.505747 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.505801 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.505812 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.505829 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.505841 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.513935 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.528310 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.540481 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.552866 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.576577 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.587850 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.598228 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.607810 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.608014 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.608174 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.608292 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.608374 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.610212 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.619844 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.630875 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.640169 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.652743 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.711349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.711654 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.711950 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.712079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.712237 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.814876 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.814952 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.814973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.815003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.815024 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.917518 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.917551 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.917560 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.917596 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:33 crc kubenswrapper[4696]: I1125 10:37:33.917606 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:33Z","lastTransitionTime":"2025-11-25T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.020577 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.021233 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.021254 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.021275 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.021951 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.041553 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:34 crc kubenswrapper[4696]: E1125 10:37:34.041724 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.041784 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:34 crc kubenswrapper[4696]: E1125 10:37:34.041850 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.041881 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:34 crc kubenswrapper[4696]: E1125 10:37:34.041937 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.053697 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.076235 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.086831 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.099931 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.112416 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.124646 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.124711 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.124722 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.124739 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.124750 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.126411 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.138530 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.167410 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.181395 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.196438 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.208184 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.218114 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.226635 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.226857 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.226940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.227039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.227127 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.228284 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.238410 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.250990 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.261633 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.274350 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.329098 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.329142 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.329152 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.329176 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.329187 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.431158 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.431188 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.431197 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.431210 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.431220 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.533169 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.533210 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.533221 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.533237 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.533247 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.635739 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.635819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.635841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.635865 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.635882 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.738483 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.738576 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.738600 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.738623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.738639 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.841509 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.841560 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.841573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.841591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.841609 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.944524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.944575 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.944591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.944610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:34 crc kubenswrapper[4696]: I1125 10:37:34.944643 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:34Z","lastTransitionTime":"2025-11-25T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.041837 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.041990 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.046463 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.046498 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.046508 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.046522 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.046532 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.148720 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.148766 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.148777 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.148796 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.148808 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.252348 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.252419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.252442 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.252468 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.252487 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.311724 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.326797 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.334487 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.355297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.355376 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.355395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.355422 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.355440 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.357333 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.375286 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.393204 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.403077 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.420927 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.436053 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.451138 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.458352 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.458392 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.458408 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.458429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.458446 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.469466 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.486038 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.498281 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.520287 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.534071 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.545415 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.555972 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.560579 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.560628 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.560646 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.560689 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.560703 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.567764 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.579223 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.663954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.663993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.664008 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.664026 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.664039 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.766566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.766893 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.766911 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.766927 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.766940 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.869631 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.869684 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.869698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.869714 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.869725 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.913422 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.913693 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:38:07.91362922 +0000 UTC m=+84.727245837 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.913757 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.913807 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.913843 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.913869 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.913960 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.913993 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914004 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914015 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914024 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:38:07.914005552 +0000 UTC m=+84.727622139 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914041 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:38:07.914033013 +0000 UTC m=+84.727649600 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914066 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914106 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914130 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914153 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914225 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:38:07.914202119 +0000 UTC m=+84.727818756 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:37:35 crc kubenswrapper[4696]: E1125 10:37:35.914296 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:38:07.91425348 +0000 UTC m=+84.727870157 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.971622 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.971682 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.971696 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.971712 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:35 crc kubenswrapper[4696]: I1125 10:37:35.971722 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:35Z","lastTransitionTime":"2025-11-25T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.042625 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.042696 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:36 crc kubenswrapper[4696]: E1125 10:37:36.042828 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:36 crc kubenswrapper[4696]: E1125 10:37:36.042961 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.043043 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:36 crc kubenswrapper[4696]: E1125 10:37:36.043090 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.074371 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.074543 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.074565 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.074591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.074608 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.177287 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.177337 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.177345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.177359 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.177387 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.279452 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.279491 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.279502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.279519 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.279530 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.388542 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.388639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.388694 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.388728 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.388757 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.492601 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.492647 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.492682 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.492699 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.492711 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.594413 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.594448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.594460 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.594476 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.594486 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.697162 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.697213 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.697222 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.697236 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.697246 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.799941 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.800003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.800016 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.800033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.800044 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.909081 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.909125 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.909136 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.909152 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:36 crc kubenswrapper[4696]: I1125 10:37:36.909163 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:36Z","lastTransitionTime":"2025-11-25T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.011586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.011694 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.011707 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.011725 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.011737 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.041229 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:37 crc kubenswrapper[4696]: E1125 10:37:37.041381 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.114590 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.114648 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.114656 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.114687 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.114696 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.217172 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.217215 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.217226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.217242 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.217251 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.320033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.320080 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.320092 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.320107 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.320118 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.423078 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.423152 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.423177 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.423205 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.423223 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.525817 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.525854 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.525866 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.525881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.525893 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.629279 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.629345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.629370 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.629400 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.629422 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.731549 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.731594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.731604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.731620 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.731635 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.834274 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.834325 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.834337 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.834357 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.834370 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.936403 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.936443 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.936455 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.936473 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:37 crc kubenswrapper[4696]: I1125 10:37:37.936485 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:37Z","lastTransitionTime":"2025-11-25T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.038439 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.038484 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.038496 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.038515 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.038527 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.041738 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:38 crc kubenswrapper[4696]: E1125 10:37:38.041838 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.041946 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.041984 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:38 crc kubenswrapper[4696]: E1125 10:37:38.042104 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:38 crc kubenswrapper[4696]: E1125 10:37:38.042240 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.141386 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.141428 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.141437 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.141451 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.141462 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.244374 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.244422 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.244438 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.244461 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.244476 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.346711 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.346769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.346787 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.346814 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.346832 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.448983 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.449019 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.449030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.449045 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.449060 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.552755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.552826 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.552850 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.552876 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.552893 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.655769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.655834 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.655846 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.655862 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.655873 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.758889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.758950 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.758966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.758991 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.759006 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.861484 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.861541 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.861552 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.861568 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.861597 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.965386 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.965504 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.965517 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.965537 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:38 crc kubenswrapper[4696]: I1125 10:37:38.965556 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:38Z","lastTransitionTime":"2025-11-25T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.041277 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:39 crc kubenswrapper[4696]: E1125 10:37:39.041907 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.068159 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.068478 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.068615 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.068803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.069053 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.171413 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.171488 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.171502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.171539 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.171555 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.274103 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.274154 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.274163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.274179 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.274194 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.376938 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.377241 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.377311 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.377374 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.377432 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.480178 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.480218 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.480229 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.480243 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.480253 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.582249 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.582483 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.582589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.582737 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.582824 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.620630 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.634278 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.647022 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.659086 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.670379 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.680338 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.684516 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.684548 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.684559 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.684597 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.684607 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.692015 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.708392 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.718646 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.728877 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.739598 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.747932 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.757824 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.769209 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.781243 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.786018 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.786212 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.786297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.786378 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.786461 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.801351 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.812225 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.824476 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.832550 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.889243 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.889311 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.889337 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.889359 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.889375 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.991971 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.992005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.992016 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.992030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:39 crc kubenswrapper[4696]: I1125 10:37:39.992041 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:39Z","lastTransitionTime":"2025-11-25T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.041884 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.042007 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.042178 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.042224 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.042317 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.042359 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.094440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.094469 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.094479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.094490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.094499 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.197568 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.197602 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.197612 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.197629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.197641 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.296853 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.296906 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.296927 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.296956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.296977 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.317171 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.321263 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.321321 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.321333 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.321350 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.321361 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.341446 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.345695 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.345728 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.345736 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.345749 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.345760 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.359693 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.363589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.363615 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.363625 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.363641 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.363653 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.376112 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.380049 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.380098 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.380114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.380135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.380152 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.406191 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:40 crc kubenswrapper[4696]: E1125 10:37:40.406429 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.408123 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.408170 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.408186 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.408205 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.408219 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.510861 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.510889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.510899 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.510914 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.510924 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.613041 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.613115 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.613126 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.613141 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.613151 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.716596 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.716711 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.716737 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.716789 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.716818 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.820378 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.820446 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.820464 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.820491 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.820508 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.923452 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.923495 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.923514 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.923532 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:40 crc kubenswrapper[4696]: I1125 10:37:40.923543 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:40Z","lastTransitionTime":"2025-11-25T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.026268 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.026331 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.026341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.026361 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.026373 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.041686 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:41 crc kubenswrapper[4696]: E1125 10:37:41.041853 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.128859 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.128915 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.128929 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.128945 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.128957 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.230645 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.230694 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.230705 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.230718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.230726 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.333110 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.333163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.333196 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.333219 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.333232 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.436523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.436563 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.436571 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.436586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.436595 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.538332 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.538371 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.538380 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.538395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.538404 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.640489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.640524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.640534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.640549 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.640561 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.743614 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.743700 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.743715 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.743737 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.743758 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.845906 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.845958 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.845968 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.845982 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.845990 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.948346 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.948396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.948409 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.948426 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:41 crc kubenswrapper[4696]: I1125 10:37:41.948437 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:41Z","lastTransitionTime":"2025-11-25T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.041287 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.041347 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.041299 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:42 crc kubenswrapper[4696]: E1125 10:37:42.041413 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:42 crc kubenswrapper[4696]: E1125 10:37:42.041517 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:42 crc kubenswrapper[4696]: E1125 10:37:42.041576 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.049881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.049922 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.049939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.049952 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.049962 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.152034 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.152063 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.152073 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.152090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.152099 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.254607 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.254650 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.254659 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.254698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.254712 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.357020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.357096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.357108 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.357124 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.357133 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.464790 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.464837 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.464845 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.465057 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.465069 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.568609 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.568655 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.568695 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.568719 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.568737 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.671977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.672043 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.672055 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.672079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.672094 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.774330 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.774375 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.774384 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.774397 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.774408 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.878297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.878365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.878382 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.878413 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.878433 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.981946 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.982010 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.982020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.982040 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:42 crc kubenswrapper[4696]: I1125 10:37:42.982053 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:42Z","lastTransitionTime":"2025-11-25T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.041956 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:43 crc kubenswrapper[4696]: E1125 10:37:43.042133 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.084880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.084938 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.084954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.084975 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.084984 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.188039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.188090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.188104 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.188122 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.188134 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.291009 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.291073 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.291106 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.291134 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.291155 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.393375 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.393418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.393427 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.393440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.393450 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.495986 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.496021 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.496034 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.496051 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.496063 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.598208 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.598235 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.598248 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.598260 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.598268 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.700835 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.700872 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.700883 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.700900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.700911 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.803779 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.803839 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.803855 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.803880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.803903 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.906842 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.906920 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.906936 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.906962 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:43 crc kubenswrapper[4696]: I1125 10:37:43.906978 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:43Z","lastTransitionTime":"2025-11-25T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.011314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.011377 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.011389 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.011407 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.011422 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.041744 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.041819 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:44 crc kubenswrapper[4696]: E1125 10:37:44.041965 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.042015 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:44 crc kubenswrapper[4696]: E1125 10:37:44.042146 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:44 crc kubenswrapper[4696]: E1125 10:37:44.042324 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.066828 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.095084 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.113017 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.115980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.116031 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.116042 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.116058 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.116069 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.127109 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.138457 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.152736 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.167034 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.180135 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.192259 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.205631 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.218566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.218622 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.218638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.218681 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.218697 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.223996 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.236930 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.249161 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.261401 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.274524 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.285027 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.296177 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.307513 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.321168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.321214 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.321226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.321242 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.321253 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.422935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.423182 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.423269 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.423348 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.423426 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.525917 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.525962 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.525972 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.525988 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.525999 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.745815 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.745878 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.745890 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.745908 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.745919 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.848097 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.848156 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.848168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.848186 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.848202 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.950716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.950755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.950764 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.950778 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:44 crc kubenswrapper[4696]: I1125 10:37:44.950789 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:44Z","lastTransitionTime":"2025-11-25T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.041981 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:45 crc kubenswrapper[4696]: E1125 10:37:45.042140 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.058414 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.058463 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.058472 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.058487 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.058497 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.161801 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.162083 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.162094 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.162110 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.162120 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.263961 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.263997 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.264008 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.264023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.264034 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.366280 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.366318 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.366325 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.366341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.366351 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.467942 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.467973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.467982 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.467996 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.468005 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.570327 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.570394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.570414 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.570431 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.570443 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.672717 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.672749 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.672758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.672769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.672778 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.774885 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.774928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.774939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.774954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.774966 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.876463 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.876527 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.876543 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.876563 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.876578 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.978713 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.978773 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.978786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.978804 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:45 crc kubenswrapper[4696]: I1125 10:37:45.978818 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:45Z","lastTransitionTime":"2025-11-25T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.042003 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.042134 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.042309 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:46 crc kubenswrapper[4696]: E1125 10:37:46.042383 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:46 crc kubenswrapper[4696]: E1125 10:37:46.042573 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:46 crc kubenswrapper[4696]: E1125 10:37:46.042650 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.082014 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.082085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.082098 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.082117 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.082135 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.183922 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.183965 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.183984 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.184002 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.184017 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.286409 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.286449 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.286460 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.286477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.286489 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.388518 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.388557 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.388566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.388581 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.388591 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.491047 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.491109 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.491120 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.491136 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.491147 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.593880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.593954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.593968 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.593987 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.594023 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.696589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.696656 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.696682 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.696701 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.696716 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.798928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.798977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.798990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.799014 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.799029 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.901833 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.901892 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.901907 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.901928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:46 crc kubenswrapper[4696]: I1125 10:37:46.901943 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:46Z","lastTransitionTime":"2025-11-25T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.004824 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.004872 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.004889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.004909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.004923 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.041691 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:47 crc kubenswrapper[4696]: E1125 10:37:47.041866 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.106534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.106564 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.106573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.106586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.106594 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.209804 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.209866 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.209890 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.209918 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.209938 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.312359 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.312398 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.312409 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.312424 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.312435 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.414470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.414505 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.414516 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.414595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.414652 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.516923 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.516963 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.516981 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.516997 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.517009 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.621199 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.621269 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.621283 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.621304 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.621324 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.724026 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.724064 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.724073 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.724088 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.724098 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.826405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.826440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.826450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.826466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.826477 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.928440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.928485 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.928496 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.928513 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:47 crc kubenswrapper[4696]: I1125 10:37:47.928524 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:47Z","lastTransitionTime":"2025-11-25T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.030603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.030658 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.030687 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.030711 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.030720 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.041480 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.041530 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:48 crc kubenswrapper[4696]: E1125 10:37:48.041597 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.041495 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:48 crc kubenswrapper[4696]: E1125 10:37:48.041950 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:48 crc kubenswrapper[4696]: E1125 10:37:48.042000 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.042212 4696 scope.go:117] "RemoveContainer" containerID="c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd" Nov 25 10:37:48 crc kubenswrapper[4696]: E1125 10:37:48.042462 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.133482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.133520 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.133531 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.133548 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.133559 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.236532 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.236603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.236615 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.236637 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.236649 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.338907 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.338957 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.338968 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.338985 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.339378 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.442058 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.442097 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.442107 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.442121 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.442131 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.544881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.544928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.544939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.544958 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.544974 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.646997 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.647056 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.647067 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.647085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.647098 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.749256 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.749302 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.749310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.749326 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.749336 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.851802 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.851841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.851853 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.851868 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.851877 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.954832 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.954885 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.954897 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.954913 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:48 crc kubenswrapper[4696]: I1125 10:37:48.954925 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:48Z","lastTransitionTime":"2025-11-25T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.041781 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:49 crc kubenswrapper[4696]: E1125 10:37:49.041944 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.057020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.057235 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.057328 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.057419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.057500 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.146836 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:49 crc kubenswrapper[4696]: E1125 10:37:49.147005 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:49 crc kubenswrapper[4696]: E1125 10:37:49.147063 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:38:21.147045071 +0000 UTC m=+97.960661658 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.159103 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.159130 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.159139 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.159154 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.159164 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.261351 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.261386 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.261397 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.261412 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.261424 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.363380 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.363424 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.363433 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.363448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.363457 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.466727 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.466801 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.466814 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.466829 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.466841 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.568746 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.568810 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.568823 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.568840 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.568852 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.671755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.671795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.671809 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.671825 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.671835 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.773815 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.773857 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.773869 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.773883 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.773896 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.876212 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.876237 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.876245 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.876258 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.876267 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.980967 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.981030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.981042 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.981078 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:49 crc kubenswrapper[4696]: I1125 10:37:49.981091 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:49Z","lastTransitionTime":"2025-11-25T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.041450 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.041492 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.041511 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.041656 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.041897 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.041961 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.083860 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.083923 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.083935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.083969 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.083982 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.186432 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.186474 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.186486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.186502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.186514 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.288649 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.288704 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.288716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.288732 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.288743 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.391788 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.391837 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.391847 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.391860 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.391874 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.443796 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.443854 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.443864 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.443879 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.443889 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.458520 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.462401 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.462429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.462440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.462492 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.462507 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.477985 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.482161 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.482307 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.482397 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.482485 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.482568 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.496522 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.502326 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.502363 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.502375 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.502391 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.502402 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.516770 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.521027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.521079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.521089 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.521113 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.521125 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.534777 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:50 crc kubenswrapper[4696]: E1125 10:37:50.534916 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.536822 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.536866 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.536881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.536901 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.536914 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.639939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.639976 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.639988 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.640005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.640017 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.742289 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.742329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.742341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.742357 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.742370 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.844414 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.844450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.844461 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.844475 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.844486 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.946987 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.947027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.947039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.947054 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:50 crc kubenswrapper[4696]: I1125 10:37:50.947065 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:50Z","lastTransitionTime":"2025-11-25T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.041478 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:51 crc kubenswrapper[4696]: E1125 10:37:51.041726 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.048977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.049023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.049030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.049047 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.049058 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.151329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.151368 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.151377 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.151392 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.151401 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.253831 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.253877 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.253889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.253906 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.253920 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.356385 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.356427 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.356438 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.356455 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.356466 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.458442 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.458481 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.458497 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.458513 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.458524 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.560114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.560150 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.560159 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.560188 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.560198 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.662466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.662503 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.662514 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.662529 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.662541 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.764345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.764377 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.764389 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.764406 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.764418 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.866387 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.866419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.866429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.866443 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.866453 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.969545 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.969577 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.969586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.969604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:51 crc kubenswrapper[4696]: I1125 10:37:51.969614 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:51Z","lastTransitionTime":"2025-11-25T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.041780 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.041798 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.041804 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:52 crc kubenswrapper[4696]: E1125 10:37:52.041913 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:52 crc kubenswrapper[4696]: E1125 10:37:52.042138 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:52 crc kubenswrapper[4696]: E1125 10:37:52.042291 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.072333 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.072372 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.072381 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.072395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.072405 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.174681 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.174718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.174729 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.174744 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.174753 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.276392 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.276421 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.276428 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.276441 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.276450 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.377944 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.377981 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.377993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.378008 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.378019 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.478704 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/0.log" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.478752 4696 generic.go:334] "Generic (PLEG): container finished" podID="bcb29cb9-408c-48c2-bdca-2c3dce47b991" containerID="1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0" exitCode=1 Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.478785 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerDied","Data":"1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.479164 4696 scope.go:117] "RemoveContainer" containerID="1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.479285 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.479317 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.479328 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.479344 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.479356 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.499856 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.515564 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.533059 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.544786 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.556931 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.579327 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.582093 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.582226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.582319 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.582406 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.582471 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.594197 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.606715 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.622103 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.634324 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.644327 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.656348 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.668386 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.682057 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.684541 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.684576 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.684585 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.684599 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.684607 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.694576 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.707847 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.718748 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.731763 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:52Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.786497 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.786544 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.786553 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.786566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.786575 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.888500 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.888565 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.888576 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.888591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.888601 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.990981 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.991004 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.991011 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.991023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:52 crc kubenswrapper[4696]: I1125 10:37:52.991032 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:52Z","lastTransitionTime":"2025-11-25T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.042185 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:53 crc kubenswrapper[4696]: E1125 10:37:53.042315 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.093084 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.093112 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.093121 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.093135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.093145 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.195207 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.195274 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.195287 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.195306 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.195317 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.297447 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.297490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.297508 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.297524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.297535 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.399524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.399561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.399577 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.399591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.399600 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.484397 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/0.log" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.484463 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerStarted","Data":"227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.502346 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.502427 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.502440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.502460 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.502474 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.506556 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.521013 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.539710 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.553091 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.566789 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.581489 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.603333 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.604941 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.604981 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.604998 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.605019 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.605036 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.619388 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.632865 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.646584 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.658009 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.667327 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.677481 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.687907 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.700069 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.707333 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.707367 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.707377 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.707394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.707405 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.710985 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.721061 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.731972 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:53Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.809861 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.810127 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.810190 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.810255 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.810320 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.912453 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.912494 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.912505 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.912523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:53 crc kubenswrapper[4696]: I1125 10:37:53.912535 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:53Z","lastTransitionTime":"2025-11-25T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.014526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.014584 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.014596 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.014613 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.014625 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.042383 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.042422 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.042446 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:54 crc kubenswrapper[4696]: E1125 10:37:54.042532 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:54 crc kubenswrapper[4696]: E1125 10:37:54.042593 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:54 crc kubenswrapper[4696]: E1125 10:37:54.042644 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.058055 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.076037 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.087947 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.108560 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.116942 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.116980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.116988 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.117020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.117030 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.126564 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.140685 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.153851 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.170176 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.187534 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.202968 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.216467 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.220223 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.220255 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.220264 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.220278 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.220289 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.226379 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.238430 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.247098 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.256512 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.264342 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.272890 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.284113 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:54Z is after 2025-08-24T17:21:41Z" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.322828 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.323351 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.323448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.323524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.323583 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.425960 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.425994 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.426004 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.426020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.426030 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.528276 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.528320 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.528331 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.528351 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.528361 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.630546 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.630574 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.630581 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.630594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.630602 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.732507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.732803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.732882 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.732956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.733014 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.835263 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.835303 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.835312 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.835329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.835340 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.937163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.937203 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.937214 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.937232 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:54 crc kubenswrapper[4696]: I1125 10:37:54.937243 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:54Z","lastTransitionTime":"2025-11-25T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.040099 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.040135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.040144 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.040162 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.040171 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.041396 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:55 crc kubenswrapper[4696]: E1125 10:37:55.041603 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.142288 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.142341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.142352 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.142370 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.142383 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.244960 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.245013 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.245023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.245036 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.245046 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.347349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.347427 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.347440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.347457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.347468 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.449590 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.449658 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.449698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.449716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.449728 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.556896 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.556935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.556949 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.556967 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.556977 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.659449 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.659497 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.659509 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.659527 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.659538 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.761590 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.761651 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.761673 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.761689 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.761698 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.864107 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.864146 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.864349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.864364 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.864375 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.966927 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.966973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.966984 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.967001 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:55 crc kubenswrapper[4696]: I1125 10:37:55.967013 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:55Z","lastTransitionTime":"2025-11-25T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.041449 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.041520 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.041537 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:56 crc kubenswrapper[4696]: E1125 10:37:56.041647 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:56 crc kubenswrapper[4696]: E1125 10:37:56.041723 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:56 crc kubenswrapper[4696]: E1125 10:37:56.041778 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.068932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.068966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.068977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.068993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.069004 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.171062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.171380 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.171735 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.172052 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.172326 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.275583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.275975 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.276086 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.276176 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.276281 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.378532 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.378573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.378586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.378604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.378617 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.480763 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.480799 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.480808 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.480821 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.480830 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.583357 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.583405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.583416 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.583433 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.583443 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.685803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.685869 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.685881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.685899 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.685912 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.788477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.788819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.788909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.788993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.789081 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.892131 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.892160 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.892168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.892180 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.892192 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.994531 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.994577 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.994594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.994614 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:56 crc kubenswrapper[4696]: I1125 10:37:56.994629 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:56Z","lastTransitionTime":"2025-11-25T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.041536 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:57 crc kubenswrapper[4696]: E1125 10:37:57.041692 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.096625 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.096683 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.096700 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.096718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.096727 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.198890 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.198916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.198924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.198936 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.198957 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.300893 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.300947 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.300964 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.300984 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.300997 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.403980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.404050 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.404065 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.404083 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.404096 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.506419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.506486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.506509 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.506541 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.506565 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.609655 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.609727 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.609740 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.609758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.609770 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.712238 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.712418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.712432 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.712448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.712457 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.814716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.814763 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.814772 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.814786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.814796 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.917262 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.917297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.917307 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.917323 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:57 crc kubenswrapper[4696]: I1125 10:37:57.917334 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:57Z","lastTransitionTime":"2025-11-25T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.019609 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.019646 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.019660 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.019692 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.019706 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.041557 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.041586 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:37:58 crc kubenswrapper[4696]: E1125 10:37:58.041761 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.041776 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:37:58 crc kubenswrapper[4696]: E1125 10:37:58.041846 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:37:58 crc kubenswrapper[4696]: E1125 10:37:58.041946 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.122074 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.122112 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.122121 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.122135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.122145 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.224504 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.224603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.224640 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.224658 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.224749 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.327558 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.327606 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.327615 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.327630 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.327640 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.430751 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.430805 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.430818 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.430836 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.430851 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.533844 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.533913 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.533925 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.533948 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.533961 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.636908 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.636955 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.636963 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.636977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.636989 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.739045 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.739085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.739099 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.739115 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.739127 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.841075 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.841120 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.841130 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.841145 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.841155 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.943708 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.943758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.943784 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.943805 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:58 crc kubenswrapper[4696]: I1125 10:37:58.943824 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:58Z","lastTransitionTime":"2025-11-25T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.041201 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:37:59 crc kubenswrapper[4696]: E1125 10:37:59.041364 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.045822 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.045873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.045886 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.045906 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.045924 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.147802 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.147842 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.147852 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.147868 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.147880 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.250191 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.250230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.250239 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.250252 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.250261 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.352539 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.352575 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.352583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.352599 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.352608 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.455086 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.455132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.455143 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.455161 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.455174 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.557192 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.557257 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.557272 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.557292 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.557308 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.660263 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.660305 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.660318 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.660335 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.660348 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.762749 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.762795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.762807 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.762826 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.762837 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.865523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.865589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.865601 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.865618 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.865630 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.967863 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.968105 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.968194 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.968277 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:37:59 crc kubenswrapper[4696]: I1125 10:37:59.968338 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:37:59Z","lastTransitionTime":"2025-11-25T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.041844 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.041868 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.041912 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.042372 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.042457 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.042247 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.070956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.071029 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.071043 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.071062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.071076 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.173724 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.173793 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.173819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.173851 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.173886 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.276083 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.276126 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.276134 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.276150 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.276160 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.378087 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.378112 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.378119 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.378132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.378140 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.480489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.480567 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.480579 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.480596 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.480604 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.583082 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.583119 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.583128 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.583140 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.583150 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.685876 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.685925 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.685936 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.685954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.685964 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.716591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.716622 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.716632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.716645 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.716655 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.728403 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.732799 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.732841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.732856 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.732875 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.732893 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.746318 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.750094 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.750162 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.750180 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.750197 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.750212 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.762723 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.773574 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.773602 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.773610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.773623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.773633 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.784752 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.787715 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.787765 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.787780 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.787800 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.787815 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.799096 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:00 crc kubenswrapper[4696]: E1125 10:38:00.799247 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.800719 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.800761 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.800775 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.800791 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.800801 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.902719 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.902789 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.902801 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.902817 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:00 crc kubenswrapper[4696]: I1125 10:38:00.902828 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:00Z","lastTransitionTime":"2025-11-25T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.005521 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.005568 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.005579 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.005595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.005608 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.041475 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:01 crc kubenswrapper[4696]: E1125 10:38:01.041655 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.107523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.107555 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.107566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.107580 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.107589 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.210314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.210365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.210377 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.210395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.210408 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.313127 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.313184 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.313199 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.313226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.313241 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.415954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.415993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.416006 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.416023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.416035 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.518608 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.518650 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.518680 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.518698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.518707 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.620734 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.620825 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.620833 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.620848 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.620858 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.723152 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.723182 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.723194 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.723209 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.723221 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.825989 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.826027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.826039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.826054 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.826065 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.927880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.927924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.927936 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.927955 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:01 crc kubenswrapper[4696]: I1125 10:38:01.927967 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:01Z","lastTransitionTime":"2025-11-25T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.030248 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.030284 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.030295 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.030310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.030321 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.041995 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:02 crc kubenswrapper[4696]: E1125 10:38:02.042152 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.042193 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:02 crc kubenswrapper[4696]: E1125 10:38:02.042343 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.042485 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:02 crc kubenswrapper[4696]: E1125 10:38:02.042595 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.133297 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.133336 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.133348 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.133362 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.133371 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.236169 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.236219 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.236228 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.236243 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.236252 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.338630 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.338689 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.338703 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.338718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.338729 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.440837 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.440865 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.440873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.440889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.440900 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.543916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.544521 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.544614 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.544730 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.544869 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.647914 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.648174 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.648267 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.648386 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.648466 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.751156 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.751227 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.751240 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.751260 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.751273 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.854059 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.854118 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.854132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.854153 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.854167 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.957216 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.957261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.957296 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.957330 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:02 crc kubenswrapper[4696]: I1125 10:38:02.957342 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:02Z","lastTransitionTime":"2025-11-25T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.041952 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:03 crc kubenswrapper[4696]: E1125 10:38:03.042087 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.042290 4696 scope.go:117] "RemoveContainer" containerID="c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.059718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.059759 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.059767 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.059781 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.059791 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.162060 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.162114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.162126 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.162143 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.162158 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.264714 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.264751 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.264786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.264803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.264813 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.367259 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.367329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.367340 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.367395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.367412 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.469687 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.469717 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.469726 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.469739 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.469766 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.515978 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/2.log" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.518253 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.519515 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.533476 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.545562 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.556947 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.569612 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.571953 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.571994 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.572007 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.572022 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.572033 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.586419 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.598691 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.613449 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.624987 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.637943 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.653531 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.663359 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.672998 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.674557 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.674598 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.674610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.674628 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.674642 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.684584 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.696531 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.721787 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.737634 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.748988 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.773106 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.779419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.779450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.779459 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.779471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.779480 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.882267 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.882308 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.882320 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.882336 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.882347 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.985092 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.985129 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.985139 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.985154 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:03 crc kubenswrapper[4696]: I1125 10:38:03.985164 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:03Z","lastTransitionTime":"2025-11-25T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.041856 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.041890 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.041926 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:04 crc kubenswrapper[4696]: E1125 10:38:04.042264 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:04 crc kubenswrapper[4696]: E1125 10:38:04.042448 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:04 crc kubenswrapper[4696]: E1125 10:38:04.042529 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.059949 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.071387 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.087646 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.087708 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.087723 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.087741 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.087752 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.090978 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.109411 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.123198 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.135611 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.146814 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.160543 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.181324 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.189682 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.189724 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.189733 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.189749 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.189759 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.196040 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.207279 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.217078 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.226752 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.235869 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.245818 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.257743 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.275541 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.289344 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.291928 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.291977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.291989 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.292007 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.292018 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.394362 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.394400 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.394412 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.394427 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.394436 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.496482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.496741 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.496805 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.496884 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.496974 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.523747 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/3.log" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.524875 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/2.log" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.527632 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" exitCode=1 Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.527685 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.527867 4696 scope.go:117] "RemoveContainer" containerID="c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.528442 4696 scope.go:117] "RemoveContainer" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" Nov 25 10:38:04 crc kubenswrapper[4696]: E1125 10:38:04.528675 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.540795 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.560151 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.571858 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.586067 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.596299 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.599142 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.599179 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.599191 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.599205 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.599216 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.609600 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.622004 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.635063 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.646872 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.658835 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.670683 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.688860 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3db9956ce30dce8796210f23e4813f51d932e52da0054962b8d16ceb86074dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:31Z\\\",\\\"message\\\":\\\"_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:37:31.813778 6270 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-vfbbz\\\\nF1125 10:37:31.814594 6270 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:37:31Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:37:31.814590 6270 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balance\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:38:04Z\\\",\\\"message\\\":\\\"l:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892596 6632 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:38:03.892600 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892626 6632 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:38:03.892627 6632 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 10:38:03.892700 6632 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.700342 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.701596 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.701641 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.701653 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.701692 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.701703 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.712267 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.723869 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.734147 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.752179 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.762500 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:04Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.803831 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.803877 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.803887 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.803900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.803915 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.906218 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.906253 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.906263 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.906277 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:04 crc kubenswrapper[4696]: I1125 10:38:04.906285 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:04Z","lastTransitionTime":"2025-11-25T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.009337 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.009374 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.009384 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.009400 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.009412 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.041204 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:05 crc kubenswrapper[4696]: E1125 10:38:05.041357 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.112208 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.112260 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.112272 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.112288 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.112298 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.214965 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.215003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.215013 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.215050 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.215061 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.318314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.318369 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.318379 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.318396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.318407 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.420401 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.420440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.420451 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.420466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.420479 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.522132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.522178 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.522186 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.522200 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.522210 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.531400 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/3.log" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.534834 4696 scope.go:117] "RemoveContainer" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" Nov 25 10:38:05 crc kubenswrapper[4696]: E1125 10:38:05.535015 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.545170 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.564878 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.576549 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.587316 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.598332 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.610740 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.624986 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.625047 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.625057 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.625071 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.625089 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.629603 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.641913 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.656062 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.666167 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.679262 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.697779 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:38:04Z\\\",\\\"message\\\":\\\"l:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892596 6632 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:38:03.892600 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892626 6632 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:38:03.892627 6632 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 10:38:03.892700 6632 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:38:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.710800 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.722101 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.727078 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.727122 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.727133 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.727148 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.727157 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.734716 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.746858 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.757104 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.768330 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:05Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.829320 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.829357 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.829367 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.829380 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.829390 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.932513 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.932873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.932885 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.932945 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:05 crc kubenswrapper[4696]: I1125 10:38:05.932958 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:05Z","lastTransitionTime":"2025-11-25T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.035674 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.035705 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.035715 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.035729 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.035737 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.041617 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.041623 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:06 crc kubenswrapper[4696]: E1125 10:38:06.041778 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:06 crc kubenswrapper[4696]: E1125 10:38:06.041881 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.041640 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:06 crc kubenswrapper[4696]: E1125 10:38:06.041955 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.138449 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.138507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.138518 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.138556 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.138570 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.242168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.242246 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.242265 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.242292 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.242305 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.344606 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.344655 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.344702 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.344722 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.344732 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.447593 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.447628 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.447678 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.447697 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.447710 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.550283 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.550319 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.550329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.550342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.550352 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.652932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.652969 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.652979 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.652992 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.653001 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.756726 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.756781 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.756792 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.756808 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.756817 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.858800 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.858836 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.858845 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.858857 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.858865 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.960361 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.960395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.960405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.960421 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:06 crc kubenswrapper[4696]: I1125 10:38:06.960435 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:06Z","lastTransitionTime":"2025-11-25T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.041882 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.042059 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.062482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.062522 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.062533 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.062547 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.062560 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.165534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.165569 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.165580 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.165594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.165605 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.268345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.268623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.268761 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.268877 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.268977 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.371364 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.371419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.371430 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.371443 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.371452 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.475358 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.475396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.475408 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.475429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.475445 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.579081 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.579516 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.579617 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.579732 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.579812 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.682069 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.682702 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.682789 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.682894 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.682989 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.786552 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.786854 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.786930 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.787025 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.787109 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.889130 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.889210 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.889221 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.889242 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.889253 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.934800 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.934893 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.934930 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.934950 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935003 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935003 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.934981432 +0000 UTC m=+148.748598029 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.935049 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935079 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.935069484 +0000 UTC m=+148.748686071 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935107 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935123 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935137 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935167 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.935157557 +0000 UTC m=+148.748774144 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935217 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935243 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.935235539 +0000 UTC m=+148.748852126 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935289 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935301 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935309 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:38:07 crc kubenswrapper[4696]: E1125 10:38:07.935328 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.935322662 +0000 UTC m=+148.748939249 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.991708 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.991753 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.991762 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.991777 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:07 crc kubenswrapper[4696]: I1125 10:38:07.991786 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:07Z","lastTransitionTime":"2025-11-25T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.042193 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.042268 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:08 crc kubenswrapper[4696]: E1125 10:38:08.042336 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:08 crc kubenswrapper[4696]: E1125 10:38:08.042424 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.042193 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:08 crc kubenswrapper[4696]: E1125 10:38:08.042540 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.094419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.094459 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.094470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.094486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.094495 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.196464 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.196506 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.196517 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.196541 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.196553 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.298626 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.298679 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.298688 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.298703 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.298712 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.401657 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.401717 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.401729 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.401746 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.401756 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.503896 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.503926 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.503935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.503950 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.503961 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.606423 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.606470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.606478 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.606494 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.606506 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.710078 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.710862 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.710898 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.710922 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.710937 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.813880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.813966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.813982 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.814012 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.814029 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.916947 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.916987 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.917000 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.917016 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:08 crc kubenswrapper[4696]: I1125 10:38:08.917027 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:08Z","lastTransitionTime":"2025-11-25T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.020394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.020451 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.020489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.020511 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.020524 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.041730 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:09 crc kubenswrapper[4696]: E1125 10:38:09.041894 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.123247 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.123318 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.123331 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.123350 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.123362 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.226552 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.226603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.226615 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.226632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.226649 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.329214 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.329261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.329272 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.329287 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.329301 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.431586 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.431623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.431632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.431645 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.431657 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.533490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.533524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.533535 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.533550 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.533561 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.635899 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.635940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.635952 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.635967 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.635979 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.738520 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.738556 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.738566 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.738583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.738593 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.841550 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.841595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.841605 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.841621 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.841631 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.944082 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.944138 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.944149 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.944171 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:09 crc kubenswrapper[4696]: I1125 10:38:09.944183 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:09Z","lastTransitionTime":"2025-11-25T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.041951 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.042012 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:10 crc kubenswrapper[4696]: E1125 10:38:10.042095 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.041974 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:10 crc kubenswrapper[4696]: E1125 10:38:10.042271 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:10 crc kubenswrapper[4696]: E1125 10:38:10.043313 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.052012 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.052054 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.052064 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.052080 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.052090 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.154764 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.154808 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.154820 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.154847 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.154862 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.256739 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.257006 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.257109 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.257196 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.257280 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.359542 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.359580 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.359591 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.359608 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.359620 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.462392 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.462430 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.462442 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.462458 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.462469 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.564980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.565079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.565096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.565116 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.565131 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.667967 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.668031 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.668044 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.668061 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.668074 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.770561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.770607 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.770623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.770643 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.770657 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.873634 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.873728 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.873745 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.873769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.873787 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.976239 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.976284 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.976296 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.976313 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:10 crc kubenswrapper[4696]: I1125 10:38:10.976324 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:10Z","lastTransitionTime":"2025-11-25T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.041634 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.041773 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.078570 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.078629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.078650 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.078713 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.078736 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.181035 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.181091 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.181110 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.181132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.181148 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.182518 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.182582 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.182595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.182611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.182622 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.196509 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.200003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.200069 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.200085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.200114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.200130 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.214189 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.218048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.218073 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.218082 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.218096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.218105 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.230741 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.233982 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.234035 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.234052 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.234070 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.234081 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.245492 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.248948 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.249032 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.249048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.249072 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.249090 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.262944 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:11Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:11 crc kubenswrapper[4696]: E1125 10:38:11.263107 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.283841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.283898 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.283911 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.283931 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.283946 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.386339 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.386385 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.386396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.386412 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.386423 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.489028 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.489074 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.489085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.489109 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.489120 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.591610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.591681 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.591735 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.591756 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.591770 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.695243 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.695303 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.695314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.695342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.695352 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.798188 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.798233 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.798245 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.798264 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.798276 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.901268 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.901304 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.901314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.901330 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:11 crc kubenswrapper[4696]: I1125 10:38:11.901341 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:11Z","lastTransitionTime":"2025-11-25T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.004353 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.004433 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.004457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.004493 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.004529 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.041775 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.041822 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:12 crc kubenswrapper[4696]: E1125 10:38:12.041929 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.041993 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:12 crc kubenswrapper[4696]: E1125 10:38:12.042052 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:12 crc kubenswrapper[4696]: E1125 10:38:12.042176 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.107279 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.107317 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.107327 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.107341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.107352 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.210134 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.210171 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.210180 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.210195 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.210205 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.313477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.313533 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.313556 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.313578 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.313594 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.416042 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.416089 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.416103 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.416121 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.416133 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.518055 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.518098 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.518107 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.518122 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.518132 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.620434 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.620490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.620499 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.620514 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.620523 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.723401 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.723438 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.723451 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.723467 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.723479 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.826773 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.826871 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.826889 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.826909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.826924 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.930253 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.930323 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.930349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.930379 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:12 crc kubenswrapper[4696]: I1125 10:38:12.930403 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:12Z","lastTransitionTime":"2025-11-25T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.033183 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.033233 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.033246 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.033264 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.033274 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.041482 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:13 crc kubenswrapper[4696]: E1125 10:38:13.041693 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.135218 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.135268 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.135286 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.135310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.135326 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.237944 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.238062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.238127 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.238162 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.238233 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.341336 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.341402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.341482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.341512 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.341536 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.444379 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.444436 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.444450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.444469 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.444482 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.547274 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.547322 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.547330 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.547345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.547355 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.649871 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.649925 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.649937 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.649954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.649968 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.751845 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.751881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.751894 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.751910 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.751921 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.854363 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.854400 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.854411 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.854428 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.854437 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.956764 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.956807 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.956815 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.956831 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:13 crc kubenswrapper[4696]: I1125 10:38:13.956851 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:13Z","lastTransitionTime":"2025-11-25T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.042247 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.042288 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:14 crc kubenswrapper[4696]: E1125 10:38:14.042459 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.042492 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:14 crc kubenswrapper[4696]: E1125 10:38:14.042561 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:14 crc kubenswrapper[4696]: E1125 10:38:14.042635 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.058741 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.059632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.059737 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.059764 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.059794 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.060077 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.071959 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.085914 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.100648 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.111096 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.125512 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.136683 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.150207 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.161533 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.161795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.161822 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.161833 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.161848 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.161859 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.180867 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.192938 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.203953 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.213493 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.223053 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.238283 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:38:04Z\\\",\\\"message\\\":\\\"l:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892596 6632 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:38:03.892600 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892626 6632 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:38:03.892627 6632 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 10:38:03.892700 6632 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:38:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.249987 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.261197 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.264038 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.264079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.264090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.264106 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.264117 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.271523 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:14Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.365689 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.365723 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.365733 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.365747 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.365757 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.468094 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.468389 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.468613 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.468797 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.468940 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.571088 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.571119 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.571127 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.571139 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.571147 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.673605 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.673916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.674041 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.674173 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.674293 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.777488 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.777806 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.777877 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.777943 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.778014 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.880617 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.880742 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.880755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.880771 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.880781 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.983629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.983718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.983737 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.983761 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:14 crc kubenswrapper[4696]: I1125 10:38:14.983778 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:14Z","lastTransitionTime":"2025-11-25T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.042219 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:15 crc kubenswrapper[4696]: E1125 10:38:15.042381 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.086702 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.086759 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.086777 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.086804 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.086825 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.189146 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.189201 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.189209 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.189225 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.189235 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.291992 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.292024 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.292033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.292048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.292057 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.396439 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.396476 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.396486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.396508 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.396520 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.499147 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.499208 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.499219 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.499233 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.499242 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.601458 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.601519 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.601530 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.601546 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.601557 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.704277 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.704352 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.704365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.704381 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.704392 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.807461 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.807538 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.807563 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.807588 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.807609 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.910714 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.910756 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.910767 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.910790 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:15 crc kubenswrapper[4696]: I1125 10:38:15.910802 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:15Z","lastTransitionTime":"2025-11-25T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.013346 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.013384 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.013396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.013413 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.013424 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.042257 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.042269 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:16 crc kubenswrapper[4696]: E1125 10:38:16.042398 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.042289 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:16 crc kubenswrapper[4696]: E1125 10:38:16.042482 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:16 crc kubenswrapper[4696]: E1125 10:38:16.042552 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.116594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.116632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.116644 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.116659 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.116716 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.220311 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.220352 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.220360 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.220375 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.220385 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.322649 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.322700 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.322709 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.322724 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.322733 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.425087 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.425114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.425122 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.425135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.425146 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.527638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.527693 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.527704 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.527721 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.527731 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.629878 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.629919 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.629929 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.629944 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.629956 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.732569 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.732599 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.732607 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.732620 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.732631 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.835071 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.835119 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.835135 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.835155 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.835170 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.937941 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.937990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.938010 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.938037 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:16 crc kubenswrapper[4696]: I1125 10:38:16.938054 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:16Z","lastTransitionTime":"2025-11-25T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.040551 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.040594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.040603 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.040622 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.040632 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.041194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:17 crc kubenswrapper[4696]: E1125 10:38:17.041347 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.143161 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.143212 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.143230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.143250 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.143263 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.245414 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.245473 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.245485 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.245504 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.245520 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.348720 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.348756 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.348767 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.348785 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.348797 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.451402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.451706 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.451790 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.451879 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.451962 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.553854 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.554222 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.554322 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.554398 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.554470 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.657403 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.657739 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.657905 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.658050 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.658116 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.760911 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.760968 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.760985 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.761013 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.761033 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.863871 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.863911 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.863920 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.863934 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.863943 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.965709 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.965769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.965786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.965810 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:17 crc kubenswrapper[4696]: I1125 10:38:17.965834 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:17Z","lastTransitionTime":"2025-11-25T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.041940 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:18 crc kubenswrapper[4696]: E1125 10:38:18.042090 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.042192 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.042245 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:18 crc kubenswrapper[4696]: E1125 10:38:18.042409 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:18 crc kubenswrapper[4696]: E1125 10:38:18.042572 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.057575 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.067977 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.068017 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.068027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.068041 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.068049 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.170501 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.170561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.170571 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.170584 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.170594 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.273355 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.273394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.273403 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.273418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.273429 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.376144 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.376186 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.376197 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.376217 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.376228 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.478158 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.478198 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.478209 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.478226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.478237 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.580554 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.580594 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.580602 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.580617 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.580627 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.683072 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.683103 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.683114 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.683129 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.683139 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.785809 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.786155 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.786167 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.786184 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.786194 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.888714 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.888795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.888840 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.888859 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.888872 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.990978 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.991039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.991049 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.991072 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:18 crc kubenswrapper[4696]: I1125 10:38:18.991084 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:18Z","lastTransitionTime":"2025-11-25T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.041748 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:19 crc kubenswrapper[4696]: E1125 10:38:19.041956 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.093956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.094305 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.094454 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.094567 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.094669 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.197055 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.197091 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.197102 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.197118 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.197131 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.299342 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.299399 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.299411 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.299428 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.299466 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.402091 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.402421 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.402492 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.402570 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.402631 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.504960 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.504990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.505017 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.505033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.505042 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.607453 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.607490 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.607501 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.607515 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.607526 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.710304 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.710353 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.710364 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.710382 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.710393 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.812901 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.812994 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.813023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.813058 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.813102 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.916507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.916554 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.916567 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.916584 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:19 crc kubenswrapper[4696]: I1125 10:38:19.916596 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:19Z","lastTransitionTime":"2025-11-25T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.020429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.020488 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.020500 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.020517 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.020529 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.042632 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.042647 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:20 crc kubenswrapper[4696]: E1125 10:38:20.043439 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:20 crc kubenswrapper[4696]: E1125 10:38:20.043527 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.043629 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:20 crc kubenswrapper[4696]: E1125 10:38:20.043910 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.123118 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.123158 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.123168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.123185 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.123197 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.225773 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.225798 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.225805 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.225818 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.225843 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.328302 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.328333 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.328341 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.328354 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.328735 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.431610 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.431688 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.431701 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.431716 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.431751 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.533789 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.533832 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.533842 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.533856 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.533867 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.637226 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.637290 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.637300 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.637314 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.637332 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.740500 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.740561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.740577 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.740602 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.740619 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.843252 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.843294 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.843306 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.843321 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.843332 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.946794 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.946857 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.946875 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.946899 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:20 crc kubenswrapper[4696]: I1125 10:38:20.946918 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:20Z","lastTransitionTime":"2025-11-25T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.041945 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.042233 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.042991 4696 scope.go:117] "RemoveContainer" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.043194 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.049228 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.049270 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.049286 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.049306 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.049321 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.151851 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.151972 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.151987 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.152003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.152014 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.184718 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.185045 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.185141 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs podName:19727774-75c0-47f9-9157-75b1116d7daa nodeName:}" failed. No retries permitted until 2025-11-25 10:39:25.185121479 +0000 UTC m=+161.998738056 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs") pod "network-metrics-daemon-w7l9g" (UID: "19727774-75c0-47f9-9157-75b1116d7daa") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.254786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.254822 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.254830 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.254843 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.254852 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.357021 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.357051 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.357060 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.357076 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.357088 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.459991 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.460332 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.460534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.460785 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.460976 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.563769 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.563813 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.563821 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.563834 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.563844 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.657248 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.657321 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.657334 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.657355 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.657368 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.672975 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.676992 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.677030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.677055 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.677071 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.677078 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.694689 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.698264 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.698287 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.698296 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.698310 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.698320 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.708454 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.711313 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.711361 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.711373 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.711390 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.711399 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.724183 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.727431 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.727455 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.727466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.727479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.727487 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.737326 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:38:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8849372d-5f88-4267-ab96-46be61e7678f\\\",\\\"systemUUID\\\":\\\"25dbb5c2-d084-42b3-95e0-6e02dc598ddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:21 crc kubenswrapper[4696]: E1125 10:38:21.737437 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.739007 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.739057 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.739068 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.739085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.739096 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.841611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.841909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.841999 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.842090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.842171 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.945214 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.945475 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.945555 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.945635 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:21 crc kubenswrapper[4696]: I1125 10:38:21.945749 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:21Z","lastTransitionTime":"2025-11-25T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.041350 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.041509 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.041740 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:22 crc kubenswrapper[4696]: E1125 10:38:22.041721 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:22 crc kubenswrapper[4696]: E1125 10:38:22.041908 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:22 crc kubenswrapper[4696]: E1125 10:38:22.042102 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.048439 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.048502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.048523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.048608 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.048627 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.151275 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.151354 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.151390 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.151417 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.151438 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.253466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.253595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.253627 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.253655 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.253709 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.356515 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.356558 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.356567 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.356582 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.356592 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.459436 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.459487 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.459503 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.459526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.459545 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.561754 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.561798 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.561809 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.561823 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.561835 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.664434 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.664479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.664489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.664504 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.664514 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.767167 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.767215 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.767231 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.767252 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.767266 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.869467 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.869529 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.869549 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.869573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.869588 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.972017 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.972053 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.972064 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.972090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:22 crc kubenswrapper[4696]: I1125 10:38:22.972102 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:22Z","lastTransitionTime":"2025-11-25T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.041602 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:23 crc kubenswrapper[4696]: E1125 10:38:23.041771 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.074728 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.074762 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.074771 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.074784 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.074793 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.177039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.177096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.177110 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.177131 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.177145 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.279581 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.279927 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.280016 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.280036 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.280053 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.382202 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.382244 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.382252 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.382267 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.382276 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.484563 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.484619 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.484641 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.484739 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.484763 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.586864 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.586905 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.586916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.586932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.586942 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.689118 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.689156 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.689168 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.689365 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.689397 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.792029 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.792084 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.792106 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.792129 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.792148 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.895470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.895517 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.895534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.895554 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.895568 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.998174 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.998221 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.998229 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.998243 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:23 crc kubenswrapper[4696]: I1125 10:38:23.998251 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:23Z","lastTransitionTime":"2025-11-25T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.041997 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.042031 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:24 crc kubenswrapper[4696]: E1125 10:38:24.042143 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.042186 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:24 crc kubenswrapper[4696]: E1125 10:38:24.042241 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:24 crc kubenswrapper[4696]: E1125 10:38:24.042342 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.056809 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a97246c-a44e-4d09-a3da-181f1f097d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4ed6f1536c2b7b32f707b579fe5a1da097e0e7f289d85f47f89a772505b9820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e500e1a43ae2f0c6f932ec282aefd9d78f516ab956bff74750f562de2c34c3f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc800f166e45a7259aac6f6a9aa5e04614c537c77b4646604ccd32a541c6e6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5a46cdd56c343238dd9922a6a961d8429a501cebdd76d520807e84fe23d9b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.069907 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f1b6d65d92d29c68d0579b3bee2154263b3423944657e8c69cda2518efb9b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.080088 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vfbbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b9b7335-d993-46d0-a36a-7821280b7a91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd00f168bb9d1b346c946c64b633bb6f71587bcac3c254beeaa53279b5ced1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpqmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vfbbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.089381 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51bf30c45f15928bec14dc567e27835dfd23ef3cc1797ec166086f4f7a5bf30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67pdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2mdzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.098468 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19727774-75c0-47f9-9157-75b1116d7daa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n27ht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-w7l9g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.099729 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.099771 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.099783 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.099800 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.099811 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.112773 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.131056 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e631c14f-04d5-4e23-9759-ed7fdd7463a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac743befc0dd3e057ba1a173d8f2d98958d0334975453d70b060dbe20d2304e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68d05ab59a2f0419ac87efbb3c41ec6c272426f9271cbc82f305a41949c07f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7481716710c75163a28d59623326fac604b1c3a4839058209331705a083e6c9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a95daa49aa3e53d99e1046cf2b9d02555c8d49f9c3722d8329622624c56e0ed6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62dd4eb3f5f5e5f9ae16f5202cfa64705ebd756b97bdc86486c0067c1f8a51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c04aba1a49d7712f3b6976732eb6fbeb97d6de4429e299eee776017c2c5d0bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59dccc8b1f88e912d874a4e270faf3b3eb2a52d5553ec0ec8153711ab39db001\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84874904d160b097b59043be0368d0faeec337e59dfeb7cd149dcbd29eda9280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.143646 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.161034 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4cbf858-76fc-426e-8561-f898927b3bf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ed0251683811235d66499300386627db06634aa7400c334c3ddfabe2eda9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77191c29f54002e5af67c4723cb1fbb9be2edbfa5ffce7c465eea75f3de091dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://354fcf7e20d73084ca69aac199ffdcedb472ce8787e9519ff739e7f5cbca6e89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd55f2f0f4948abf98cb0b5788ba641d93aae1a0955188e99f34cbe55a8313e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc3fc2b4b16708cebb9945d5c7f921f1d4f47e3f97da49c5f49e2abeeea3b60d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9505d50156b056423b991da8a0b3a6489f022b8f8f2128c6a6e0dd9265b7704e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8372127e1570607e708689a017655db36665e0c16e71c3fb3bd05f06893f1e60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcjxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb7d4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.172505 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gk2rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173afc30-0ef2-4c6d-b6fa-48c7905d16dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4064713821d708e03fb48e0c1cfc75c284852269683994835c48cfeb8c3cd21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pdxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gk2rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.184066 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g77vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb29cb9-408c-48c2-bdca-2c3dce47b991\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:37:52Z\\\",\\\"message\\\":\\\"2025-11-25T10:37:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed\\\\n2025-11-25T10:37:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69d108d8-9396-4d97-b609-5f1e215b63ed to /host/opt/cni/bin/\\\\n2025-11-25T10:37:06Z [verbose] multus-daemon started\\\\n2025-11-25T10:37:06Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:37:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5fkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g77vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.197947 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6dd3387-500c-411a-b208-625603bf4dd7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0975e045a554d49fdb890e19651d9d5b50068dda48051087bdcafabdb61c774\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a594ea257fbca49941677dc1162f9e1e26d81037db23e55217d2a71f3d93e20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dacfa077ac5388c6563a9804555ebd4142653506c4c9c26f7455784c8e7d0aa5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.201704 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.201758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.201778 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.201801 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.201819 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.209944 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac87cbfc-e1e5-411c-809a-1add49e739fc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7264441bd8ecb4f8941bdca823785efd2ea998fbb88ef49f32577e274aff5a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe196bdfbebf41e6668bcdcf12d335bb94611db1591d1b0183935f9b6d15704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebe196bdfbebf41e6668bcdcf12d335bb94611db1591d1b0183935f9b6d15704\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.223887 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34438097-5858-4502-983c-3e167dea3525\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 10:37:03.080080 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 10:37:03.080268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 10:37:03.081400 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3529907815/tls.crt::/tmp/serving-cert-3529907815/tls.key\\\\\\\"\\\\nI1125 10:37:03.578811 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 10:37:03.583634 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 10:37:03.583688 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 10:37:03.583725 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 10:37:03.583732 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 10:37:03.590050 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 10:37:03.590081 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590086 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 10:37:03.590090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 10:37:03.590094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 10:37:03.590097 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 10:37:03.590100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 10:37:03.590464 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 10:37:03.592322 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:36:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:36:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.245623 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.260973 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://508966effa18491afc1622fa784ce61fdc59d6050095207c039e44cac2a64397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.281492 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9bdb71510af0d493c3e61419d7dca00e1c6b9cc8654a0ace50c60a7ee242d90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34047e92a2364e0866343229771490bcb62536eb1769da0b2a3ca4617a9e682\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.304042 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.304112 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.304124 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.304141 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.304153 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.305926 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2a63620-dc3c-467f-9cf5-d688f0e233c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:38:04Z\\\",\\\"message\\\":\\\"l:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892596 6632 ovnkube.go:599] Stopped ovnkube\\\\nI1125 10:38:03.892600 6632 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 10:38:03.892626 6632 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:38:03.892627 6632 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 10:38:03.892700 6632 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:38:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:37:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wkcm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-l8nb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.320301 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ebd497-beb3-40a3-9c6a-8ad74b514311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2675bca8cc1b282625a98658bdae46058e54edde1f8fba8feaa7fe578a7134a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52d21a13e162436b483e88f5a0f61ff4ec5f1de3f879b5cf61b54dc0d8222853\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckngr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:37:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q244k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:38:24Z is after 2025-08-24T17:21:41Z" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.406794 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.406840 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.406849 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.406867 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.406877 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.510030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.510396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.510471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.510546 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.510662 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.613562 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.613620 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.613649 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.613720 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.613745 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.716871 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.716930 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.716948 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.716980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.716998 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.819017 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.819066 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.819075 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.819088 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.819099 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.921965 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.922026 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.922048 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.922076 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:24 crc kubenswrapper[4696]: I1125 10:38:24.922097 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:24Z","lastTransitionTime":"2025-11-25T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.023891 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.023926 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.023936 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.023954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.023966 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.041709 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:25 crc kubenswrapper[4696]: E1125 10:38:25.041843 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.126015 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.126057 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.126068 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.126085 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.126097 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.228334 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.228386 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.228402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.228426 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.228443 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.330538 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.330613 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.330637 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.330703 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.330737 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.432749 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.432791 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.432802 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.432821 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.432833 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.536435 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.536471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.536482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.536501 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.536513 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.638639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.638735 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.638745 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.638786 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.638798 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.741909 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.741963 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.741981 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.742006 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.742024 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.844532 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.844587 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.844598 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.844614 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.844624 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.947305 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.947378 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.947397 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.947424 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:25 crc kubenswrapper[4696]: I1125 10:38:25.947444 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:25Z","lastTransitionTime":"2025-11-25T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.041529 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.041652 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:26 crc kubenswrapper[4696]: E1125 10:38:26.041788 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.041801 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:26 crc kubenswrapper[4696]: E1125 10:38:26.041891 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:26 crc kubenswrapper[4696]: E1125 10:38:26.041950 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.049840 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.049892 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.049903 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.049916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.049925 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.152562 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.152616 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.152629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.152648 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.152695 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.256005 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.256081 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.256104 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.256138 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.256159 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.359084 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.359151 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.359164 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.359182 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.359194 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.461450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.461553 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.461631 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.461658 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.461707 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.564534 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.564611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.564634 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.564697 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.564727 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.668308 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.668370 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.668391 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.668417 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.668436 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.770905 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.771031 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.771052 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.771078 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.771098 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.873705 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.873753 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.873763 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.873780 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.873793 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.977403 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.977805 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.977918 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.978027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:26 crc kubenswrapper[4696]: I1125 10:38:26.978125 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:26Z","lastTransitionTime":"2025-11-25T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.042096 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:27 crc kubenswrapper[4696]: E1125 10:38:27.042380 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.081295 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.081355 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.081367 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.081389 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.081404 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.185394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.185463 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.185482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.185510 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.185531 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.288560 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.288597 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.288608 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.288623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.288633 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.391338 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.391390 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.391405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.391426 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.391441 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.493827 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.493861 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.493869 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.493882 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.493891 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.596962 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.597009 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.597042 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.597061 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.597072 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.699825 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.699868 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.699876 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.699891 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.699899 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.802795 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.802840 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.802853 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.802870 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.802882 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.906137 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.906213 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.906236 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.906261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:27 crc kubenswrapper[4696]: I1125 10:38:27.906278 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:27Z","lastTransitionTime":"2025-11-25T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.008090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.008148 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.008157 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.008173 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.008182 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.041313 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.041343 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.041327 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:28 crc kubenswrapper[4696]: E1125 10:38:28.041431 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:28 crc kubenswrapper[4696]: E1125 10:38:28.041512 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:28 crc kubenswrapper[4696]: E1125 10:38:28.041598 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.110612 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.110647 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.110656 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.110689 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.110700 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.212933 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.213003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.213018 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.213039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.213055 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.315109 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.315142 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.315150 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.315162 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.315172 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.417345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.417394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.417405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.417423 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.417437 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.519467 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.519503 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.519515 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.519532 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.519543 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.622220 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.622278 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.622296 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.622318 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.622336 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.724360 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.724393 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.724404 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.724418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.724426 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.826697 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.826747 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.826756 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.826773 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.826785 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.929343 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.929402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.929415 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.929430 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:28 crc kubenswrapper[4696]: I1125 10:38:28.929442 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:28Z","lastTransitionTime":"2025-11-25T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.035276 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.035408 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.035423 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.035441 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.035457 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.041274 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:29 crc kubenswrapper[4696]: E1125 10:38:29.041617 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.138202 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.138522 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.138561 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.138583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.138598 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.241169 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.241210 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.241218 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.241232 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.241241 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.343442 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.343480 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.343489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.343502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.343510 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.446031 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.446084 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.446100 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.446116 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.446126 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.548770 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.548811 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.548819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.548832 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.548841 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.650862 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.650903 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.650911 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.650925 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.650934 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.753246 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.753280 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.753288 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.753301 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.753310 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.855062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.855120 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.855133 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.855148 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.855159 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.957127 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.957163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.957172 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.957185 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:29 crc kubenswrapper[4696]: I1125 10:38:29.957193 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:29Z","lastTransitionTime":"2025-11-25T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.041252 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:30 crc kubenswrapper[4696]: E1125 10:38:30.041373 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.041400 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:30 crc kubenswrapper[4696]: E1125 10:38:30.041486 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.041571 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:30 crc kubenswrapper[4696]: E1125 10:38:30.041679 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.058826 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.058852 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.058860 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.058873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.058881 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.160423 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.160457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.160467 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.160481 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.160489 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.262993 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.263030 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.263040 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.263055 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.263066 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.365816 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.365868 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.365880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.365899 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.365911 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.468079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.468132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.468142 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.468155 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.468166 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.570336 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.570383 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.570394 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.570410 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.570422 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.672436 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.672471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.672479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.672492 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.672501 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.774817 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.774857 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.774867 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.774882 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.774893 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.877388 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.877430 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.877440 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.877455 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.877465 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.979766 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.979846 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.979887 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.979920 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:30 crc kubenswrapper[4696]: I1125 10:38:30.979942 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:30Z","lastTransitionTime":"2025-11-25T10:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.041556 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:31 crc kubenswrapper[4696]: E1125 10:38:31.041734 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.083363 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.083439 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.083456 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.083474 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.083484 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.185827 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.185892 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.185904 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.185920 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.185931 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.287755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.287819 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.287841 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.287868 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.287889 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.389823 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.389888 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.389900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.389916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.389927 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.492125 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.492178 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.492194 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.492215 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.492227 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.594449 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.594482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.594489 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.594501 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.594511 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.696834 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.696885 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.696900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.696921 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.696938 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.799357 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.799623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.799803 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.799933 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.800025 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.902270 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.902324 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.902359 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.902381 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.902395 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.906294 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.906319 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.906329 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.906372 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.906384 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:38:31Z","lastTransitionTime":"2025-11-25T10:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.953515 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8"] Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.954018 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.956158 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.956236 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.956331 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.956836 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 10:38:31 crc kubenswrapper[4696]: I1125 10:38:31.992488 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=85.992472873 podStartE2EDuration="1m25.992472873s" podCreationTimestamp="2025-11-25 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:31.991999059 +0000 UTC m=+108.805615646" watchObservedRunningTime="2025-11-25 10:38:31.992472873 +0000 UTC m=+108.806089460" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.033311 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fb7d4" podStartSLOduration=89.033289678 podStartE2EDuration="1m29.033289678s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.01891067 +0000 UTC m=+108.832527287" watchObservedRunningTime="2025-11-25 10:38:32.033289678 +0000 UTC m=+108.846906275" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.041121 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.041183 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:32 crc kubenswrapper[4696]: E1125 10:38:32.041205 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.041325 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:32 crc kubenswrapper[4696]: E1125 10:38:32.041322 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:32 crc kubenswrapper[4696]: E1125 10:38:32.041376 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.053357 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gk2rz" podStartSLOduration=88.053344324 podStartE2EDuration="1m28.053344324s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.034964068 +0000 UTC m=+108.848580655" watchObservedRunningTime="2025-11-25 10:38:32.053344324 +0000 UTC m=+108.866960931" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.064161 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=14.064124195 podStartE2EDuration="14.064124195s" podCreationTimestamp="2025-11-25 10:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.063562958 +0000 UTC m=+108.877179555" watchObservedRunningTime="2025-11-25 10:38:32.064124195 +0000 UTC m=+108.877740782" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.064292 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.06428656 podStartE2EDuration="1m28.06428656s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.053585362 +0000 UTC m=+108.867201959" watchObservedRunningTime="2025-11-25 10:38:32.06428656 +0000 UTC m=+108.877903147" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.093313 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c17c189b-3507-4f3c-9563-d6de26c7cc83-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.093368 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c17c189b-3507-4f3c-9563-d6de26c7cc83-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.093406 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c17c189b-3507-4f3c-9563-d6de26c7cc83-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.093423 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c17c189b-3507-4f3c-9563-d6de26c7cc83-service-ca\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.093439 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c17c189b-3507-4f3c-9563-d6de26c7cc83-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.104089 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.104072334 podStartE2EDuration="1m28.104072334s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.090305304 +0000 UTC m=+108.903921891" watchObservedRunningTime="2025-11-25 10:38:32.104072334 +0000 UTC m=+108.917688921" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.177144 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-g77vb" podStartSLOduration=89.177107207 podStartE2EDuration="1m29.177107207s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.176465528 +0000 UTC m=+108.990082135" watchObservedRunningTime="2025-11-25 10:38:32.177107207 +0000 UTC m=+108.990723794" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.189434 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q244k" podStartSLOduration=88.189415004 podStartE2EDuration="1m28.189415004s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.18861739 +0000 UTC m=+109.002233977" watchObservedRunningTime="2025-11-25 10:38:32.189415004 +0000 UTC m=+109.003031581" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195178 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c17c189b-3507-4f3c-9563-d6de26c7cc83-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195255 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c17c189b-3507-4f3c-9563-d6de26c7cc83-service-ca\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c17c189b-3507-4f3c-9563-d6de26c7cc83-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c17c189b-3507-4f3c-9563-d6de26c7cc83-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195434 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c17c189b-3507-4f3c-9563-d6de26c7cc83-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195447 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c17c189b-3507-4f3c-9563-d6de26c7cc83-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.195688 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c17c189b-3507-4f3c-9563-d6de26c7cc83-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.196685 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c17c189b-3507-4f3c-9563-d6de26c7cc83-service-ca\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.210349 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c17c189b-3507-4f3c-9563-d6de26c7cc83-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.213983 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c17c189b-3507-4f3c-9563-d6de26c7cc83-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-shtg8\" (UID: \"c17c189b-3507-4f3c-9563-d6de26c7cc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.246224 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.246204373 podStartE2EDuration="57.246204373s" podCreationTimestamp="2025-11-25 10:37:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.225611651 +0000 UTC m=+109.039228228" watchObservedRunningTime="2025-11-25 10:38:32.246204373 +0000 UTC m=+109.059820990" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.257518 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-vfbbz" podStartSLOduration=89.257496459 podStartE2EDuration="1m29.257496459s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.257075516 +0000 UTC m=+109.070692103" watchObservedRunningTime="2025-11-25 10:38:32.257496459 +0000 UTC m=+109.071113046" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.269564 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.286931 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podStartSLOduration=89.286916835 podStartE2EDuration="1m29.286916835s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.27199372 +0000 UTC m=+109.085610327" watchObservedRunningTime="2025-11-25 10:38:32.286916835 +0000 UTC m=+109.100533422" Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.612396 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" event={"ID":"c17c189b-3507-4f3c-9563-d6de26c7cc83","Type":"ContainerStarted","Data":"6f7264d7195f317120d01ff3b942822e4debda7f91beb540910636723b859d44"} Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.612452 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" event={"ID":"c17c189b-3507-4f3c-9563-d6de26c7cc83","Type":"ContainerStarted","Data":"b75cf6b3bdf480c5a3cde30dce2ef55971630534e9174d4cb620bdbc253b63c1"} Nov 25 10:38:32 crc kubenswrapper[4696]: I1125 10:38:32.624788 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-shtg8" podStartSLOduration=89.624768117 podStartE2EDuration="1m29.624768117s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:32.62450857 +0000 UTC m=+109.438125157" watchObservedRunningTime="2025-11-25 10:38:32.624768117 +0000 UTC m=+109.438384704" Nov 25 10:38:33 crc kubenswrapper[4696]: I1125 10:38:33.041963 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:33 crc kubenswrapper[4696]: E1125 10:38:33.042205 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:34 crc kubenswrapper[4696]: I1125 10:38:34.041835 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:34 crc kubenswrapper[4696]: I1125 10:38:34.041853 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:34 crc kubenswrapper[4696]: I1125 10:38:34.042974 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:34 crc kubenswrapper[4696]: E1125 10:38:34.043107 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:34 crc kubenswrapper[4696]: E1125 10:38:34.043221 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:34 crc kubenswrapper[4696]: I1125 10:38:34.043316 4696 scope.go:117] "RemoveContainer" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" Nov 25 10:38:34 crc kubenswrapper[4696]: E1125 10:38:34.043445 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:34 crc kubenswrapper[4696]: E1125 10:38:34.043452 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-l8nb9_openshift-ovn-kubernetes(f2a63620-dc3c-467f-9cf5-d688f0e233c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" Nov 25 10:38:35 crc kubenswrapper[4696]: I1125 10:38:35.042003 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:35 crc kubenswrapper[4696]: E1125 10:38:35.042348 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:36 crc kubenswrapper[4696]: I1125 10:38:36.042001 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:36 crc kubenswrapper[4696]: E1125 10:38:36.042105 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:36 crc kubenswrapper[4696]: I1125 10:38:36.042136 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:36 crc kubenswrapper[4696]: I1125 10:38:36.042006 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:36 crc kubenswrapper[4696]: E1125 10:38:36.042250 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:36 crc kubenswrapper[4696]: E1125 10:38:36.042271 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:37 crc kubenswrapper[4696]: I1125 10:38:37.041727 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:37 crc kubenswrapper[4696]: E1125 10:38:37.041880 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.041765 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.041868 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.041892 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:38 crc kubenswrapper[4696]: E1125 10:38:38.042510 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:38 crc kubenswrapper[4696]: E1125 10:38:38.042624 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:38 crc kubenswrapper[4696]: E1125 10:38:38.042911 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.638952 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/1.log" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.639399 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/0.log" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.639440 4696 generic.go:334] "Generic (PLEG): container finished" podID="bcb29cb9-408c-48c2-bdca-2c3dce47b991" containerID="227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964" exitCode=1 Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.639474 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerDied","Data":"227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964"} Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.639518 4696 scope.go:117] "RemoveContainer" containerID="1ed1b66cdeed689c01cb88602549a1e95701e2ec3c69c5c7f1897b7819842df0" Nov 25 10:38:38 crc kubenswrapper[4696]: I1125 10:38:38.640001 4696 scope.go:117] "RemoveContainer" containerID="227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964" Nov 25 10:38:38 crc kubenswrapper[4696]: E1125 10:38:38.640267 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-g77vb_openshift-multus(bcb29cb9-408c-48c2-bdca-2c3dce47b991)\"" pod="openshift-multus/multus-g77vb" podUID="bcb29cb9-408c-48c2-bdca-2c3dce47b991" Nov 25 10:38:39 crc kubenswrapper[4696]: I1125 10:38:39.042135 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:39 crc kubenswrapper[4696]: E1125 10:38:39.042285 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:39 crc kubenswrapper[4696]: I1125 10:38:39.645140 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/1.log" Nov 25 10:38:40 crc kubenswrapper[4696]: I1125 10:38:40.041758 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:40 crc kubenswrapper[4696]: I1125 10:38:40.041864 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:40 crc kubenswrapper[4696]: E1125 10:38:40.041915 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:40 crc kubenswrapper[4696]: I1125 10:38:40.041962 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:40 crc kubenswrapper[4696]: E1125 10:38:40.042080 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:40 crc kubenswrapper[4696]: E1125 10:38:40.042248 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:41 crc kubenswrapper[4696]: I1125 10:38:41.041588 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:41 crc kubenswrapper[4696]: E1125 10:38:41.041850 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:42 crc kubenswrapper[4696]: I1125 10:38:42.041299 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:42 crc kubenswrapper[4696]: E1125 10:38:42.041454 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:42 crc kubenswrapper[4696]: I1125 10:38:42.041318 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:42 crc kubenswrapper[4696]: I1125 10:38:42.041319 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:42 crc kubenswrapper[4696]: E1125 10:38:42.041584 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:42 crc kubenswrapper[4696]: E1125 10:38:42.041806 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:43 crc kubenswrapper[4696]: I1125 10:38:43.042162 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:43 crc kubenswrapper[4696]: E1125 10:38:43.042777 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:44 crc kubenswrapper[4696]: E1125 10:38:44.020763 4696 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 10:38:44 crc kubenswrapper[4696]: I1125 10:38:44.041314 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:44 crc kubenswrapper[4696]: I1125 10:38:44.041328 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:44 crc kubenswrapper[4696]: I1125 10:38:44.042864 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:44 crc kubenswrapper[4696]: E1125 10:38:44.042993 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:44 crc kubenswrapper[4696]: E1125 10:38:44.043061 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:44 crc kubenswrapper[4696]: E1125 10:38:44.043160 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:44 crc kubenswrapper[4696]: E1125 10:38:44.164106 4696 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.041849 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:45 crc kubenswrapper[4696]: E1125 10:38:45.042063 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.043204 4696 scope.go:117] "RemoveContainer" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.668106 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/3.log" Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.670473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerStarted","Data":"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef"} Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.670928 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.700168 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podStartSLOduration=102.700152469 podStartE2EDuration="1m42.700152469s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:38:45.699296614 +0000 UTC m=+122.512913201" watchObservedRunningTime="2025-11-25 10:38:45.700152469 +0000 UTC m=+122.513769056" Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.858492 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-w7l9g"] Nov 25 10:38:45 crc kubenswrapper[4696]: I1125 10:38:45.858585 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:45 crc kubenswrapper[4696]: E1125 10:38:45.858689 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:46 crc kubenswrapper[4696]: I1125 10:38:46.042309 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:46 crc kubenswrapper[4696]: I1125 10:38:46.042357 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:46 crc kubenswrapper[4696]: E1125 10:38:46.042441 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:46 crc kubenswrapper[4696]: I1125 10:38:46.042497 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:46 crc kubenswrapper[4696]: E1125 10:38:46.042589 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:46 crc kubenswrapper[4696]: E1125 10:38:46.042658 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:48 crc kubenswrapper[4696]: I1125 10:38:48.041960 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:48 crc kubenswrapper[4696]: I1125 10:38:48.042012 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:48 crc kubenswrapper[4696]: E1125 10:38:48.042887 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:48 crc kubenswrapper[4696]: I1125 10:38:48.042055 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:48 crc kubenswrapper[4696]: I1125 10:38:48.042047 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:48 crc kubenswrapper[4696]: E1125 10:38:48.042973 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:48 crc kubenswrapper[4696]: E1125 10:38:48.043046 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:48 crc kubenswrapper[4696]: E1125 10:38:48.043180 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:49 crc kubenswrapper[4696]: E1125 10:38:49.164920 4696 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:38:50 crc kubenswrapper[4696]: I1125 10:38:50.042053 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:50 crc kubenswrapper[4696]: E1125 10:38:50.042201 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:50 crc kubenswrapper[4696]: I1125 10:38:50.042452 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:50 crc kubenswrapper[4696]: I1125 10:38:50.042453 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:50 crc kubenswrapper[4696]: E1125 10:38:50.042597 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:50 crc kubenswrapper[4696]: E1125 10:38:50.042726 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:50 crc kubenswrapper[4696]: I1125 10:38:50.042721 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:50 crc kubenswrapper[4696]: E1125 10:38:50.042837 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:51 crc kubenswrapper[4696]: I1125 10:38:51.041948 4696 scope.go:117] "RemoveContainer" containerID="227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964" Nov 25 10:38:51 crc kubenswrapper[4696]: I1125 10:38:51.689422 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/1.log" Nov 25 10:38:51 crc kubenswrapper[4696]: I1125 10:38:51.689746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerStarted","Data":"0e32d100ef8aa5d1698b8603daba8da5a1f983cd8740afe907b4552254b7ee52"} Nov 25 10:38:52 crc kubenswrapper[4696]: I1125 10:38:52.041626 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:52 crc kubenswrapper[4696]: I1125 10:38:52.041656 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:52 crc kubenswrapper[4696]: I1125 10:38:52.041623 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:52 crc kubenswrapper[4696]: E1125 10:38:52.041788 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:52 crc kubenswrapper[4696]: E1125 10:38:52.041856 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:52 crc kubenswrapper[4696]: E1125 10:38:52.041933 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:52 crc kubenswrapper[4696]: I1125 10:38:52.042122 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:52 crc kubenswrapper[4696]: E1125 10:38:52.042195 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:54 crc kubenswrapper[4696]: I1125 10:38:54.041286 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:54 crc kubenswrapper[4696]: I1125 10:38:54.041341 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:54 crc kubenswrapper[4696]: I1125 10:38:54.041294 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:54 crc kubenswrapper[4696]: I1125 10:38:54.041351 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:54 crc kubenswrapper[4696]: E1125 10:38:54.042703 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w7l9g" podUID="19727774-75c0-47f9-9157-75b1116d7daa" Nov 25 10:38:54 crc kubenswrapper[4696]: E1125 10:38:54.042848 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:38:54 crc kubenswrapper[4696]: E1125 10:38:54.042950 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:38:54 crc kubenswrapper[4696]: E1125 10:38:54.043093 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.041550 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.041732 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.042075 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.042097 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.045355 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.045522 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.045742 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.045984 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.045994 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 10:38:56 crc kubenswrapper[4696]: I1125 10:38:56.046214 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.570125 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.611020 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.611772 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.618208 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.619286 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.623118 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.623493 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.624163 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.634773 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.634939 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.635248 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.635324 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.636858 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.636880 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.637745 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.638995 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.639593 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.639942 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.640978 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.641288 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.648288 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.648615 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.648904 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.649115 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.650035 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.650317 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.650489 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.650586 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.650777 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.660633 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-clh68"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.661123 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.661309 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.661764 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.664208 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5g9hp"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.665801 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.665986 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.666812 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6sm87"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.667846 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.668639 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.670525 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.670775 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.673639 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sbgf7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.674296 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.675192 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.675336 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.675589 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.675856 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.675953 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.676368 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.677402 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.677832 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.678064 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.678482 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.678873 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.678992 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.679318 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.679472 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.679773 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.679883 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.680214 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.684032 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.684490 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.684981 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.687748 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2pqps"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.688516 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4z2br"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.688895 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.689226 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.689418 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.689598 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.690479 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.691035 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.691419 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.691470 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.691614 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.691829 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.692336 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.692015 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.692177 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.698015 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-k24kz"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.698503 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xkw2f"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.698843 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699153 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.692147 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.698627 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.698766 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.698892 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699188 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699249 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699301 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699349 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699379 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.699856 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.702124 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.702494 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.702658 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.702752 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.702852 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.702928 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.703024 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.703095 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.703168 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.704203 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705149 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705186 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705293 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705511 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705636 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705782 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705830 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.705917 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.706041 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.706059 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.706133 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.706227 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.707305 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708259 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-config\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708386 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7b36fad9-a6b9-4082-8861-95d4aba0efb7-machine-approver-tls\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708413 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvcf\" (UniqueName: \"kubernetes.io/projected/7b36fad9-a6b9-4082-8861-95d4aba0efb7-kube-api-access-4zvcf\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708442 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5939a67f-5765-4afd-905e-76c5a3c8c783-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708463 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-encryption-config\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708487 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708504 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-config\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708522 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxsg7\" (UniqueName: \"kubernetes.io/projected/c830e283-0231-4570-97f7-5583aa7d2c19-kube-api-access-dxsg7\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708541 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-serving-cert\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708565 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-client-ca\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708581 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5939a67f-5765-4afd-905e-76c5a3c8c783-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708596 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708614 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-client-ca\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708632 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be09268-4279-43e9-8da0-78c4cdd920cd-serving-cert\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708646 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c830e283-0231-4570-97f7-5583aa7d2c19-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708682 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdkrm\" (UniqueName: \"kubernetes.io/projected/5939a67f-5765-4afd-905e-76c5a3c8c783-kube-api-access-fdkrm\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708700 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-etcd-client\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708718 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m2gd\" (UniqueName: \"kubernetes.io/projected/50d299fa-89e9-41ba-8894-af097502b258-kube-api-access-4m2gd\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708740 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b36fad9-a6b9-4082-8861-95d4aba0efb7-auth-proxy-config\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708760 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89c24\" (UniqueName: \"kubernetes.io/projected/735300e4-0a1b-4e3f-9735-84abcf63cf4a-kube-api-access-89c24\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708801 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50d299fa-89e9-41ba-8894-af097502b258-audit-dir\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708827 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735300e4-0a1b-4e3f-9735-84abcf63cf4a-serving-cert\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708843 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b36fad9-a6b9-4082-8861-95d4aba0efb7-config\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708858 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708872 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c830e283-0231-4570-97f7-5583aa7d2c19-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708886 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-audit-policies\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.708901 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzntr\" (UniqueName: \"kubernetes.io/projected/0be09268-4279-43e9-8da0-78c4cdd920cd-kube-api-access-mzntr\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.723138 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.723767 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mjb8l"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.724075 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.724341 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.724759 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.743644 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.745278 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.750962 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.767970 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.782311 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.782359 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.783603 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.784355 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.784704 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.784962 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.789787 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.789977 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.790095 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.790226 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.790384 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.790555 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.791049 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.791551 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-rj85t"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.791931 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.791981 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7kd7r"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.792194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.792342 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.792406 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.793428 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.793709 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798033 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798309 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798427 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798329 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798619 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798730 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.798837 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.800125 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.800734 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.801440 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.804274 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.805033 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.805275 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nvzm5"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.806098 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bv2wn"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.807001 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.807761 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.808205 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.808608 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.812926 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-config\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.812968 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxsg7\" (UniqueName: \"kubernetes.io/projected/c830e283-0231-4570-97f7-5583aa7d2c19-kube-api-access-dxsg7\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.812993 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-serving-cert\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813027 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-client-ca\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813049 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5939a67f-5765-4afd-905e-76c5a3c8c783-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813068 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813088 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-client-ca\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813110 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be09268-4279-43e9-8da0-78c4cdd920cd-serving-cert\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813129 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c830e283-0231-4570-97f7-5583aa7d2c19-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813151 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdkrm\" (UniqueName: \"kubernetes.io/projected/5939a67f-5765-4afd-905e-76c5a3c8c783-kube-api-access-fdkrm\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813174 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-etcd-client\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813194 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m2gd\" (UniqueName: \"kubernetes.io/projected/50d299fa-89e9-41ba-8894-af097502b258-kube-api-access-4m2gd\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813217 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b36fad9-a6b9-4082-8861-95d4aba0efb7-auth-proxy-config\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813242 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89c24\" (UniqueName: \"kubernetes.io/projected/735300e4-0a1b-4e3f-9735-84abcf63cf4a-kube-api-access-89c24\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813281 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50d299fa-89e9-41ba-8894-af097502b258-audit-dir\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813317 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735300e4-0a1b-4e3f-9735-84abcf63cf4a-serving-cert\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813340 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b36fad9-a6b9-4082-8861-95d4aba0efb7-config\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813366 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813390 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c830e283-0231-4570-97f7-5583aa7d2c19-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-audit-policies\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813441 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzntr\" (UniqueName: \"kubernetes.io/projected/0be09268-4279-43e9-8da0-78c4cdd920cd-kube-api-access-mzntr\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-config\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813492 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7b36fad9-a6b9-4082-8861-95d4aba0efb7-machine-approver-tls\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813517 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvcf\" (UniqueName: \"kubernetes.io/projected/7b36fad9-a6b9-4082-8861-95d4aba0efb7-kube-api-access-4zvcf\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813552 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5939a67f-5765-4afd-905e-76c5a3c8c783-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813578 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-encryption-config\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.813611 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.814311 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.814484 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50d299fa-89e9-41ba-8894-af097502b258-audit-dir\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.816004 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c830e283-0231-4570-97f7-5583aa7d2c19-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.816326 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.816742 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-client-ca\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.817756 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-config\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.818946 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b36fad9-a6b9-4082-8861-95d4aba0efb7-auth-proxy-config\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.819440 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.820436 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-client-ca\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.821532 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.821787 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.822282 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.822723 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.823006 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.824497 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b36fad9-a6b9-4082-8861-95d4aba0efb7-config\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.826142 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5939a67f-5765-4afd-905e-76c5a3c8c783-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.829128 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.834525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.833647 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-config\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.840541 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5939a67f-5765-4afd-905e-76c5a3c8c783-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.841097 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50d299fa-89e9-41ba-8894-af097502b258-audit-policies\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.841530 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-whvmw"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.842454 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.842972 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7b36fad9-a6b9-4082-8861-95d4aba0efb7-machine-approver-tls\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.842983 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.847136 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-encryption-config\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.848699 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.848837 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.849583 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.865749 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkn7w"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.866534 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.866561 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h6tc4"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.867068 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.867421 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.868457 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.869582 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.882650 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c830e283-0231-4570-97f7-5583aa7d2c19-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.884071 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be09268-4279-43e9-8da0-78c4cdd920cd-serving-cert\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.885259 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.885472 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-serving-cert\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.886069 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50d299fa-89e9-41ba-8894-af097502b258-etcd-client\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.883656 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735300e4-0a1b-4e3f-9735-84abcf63cf4a-serving-cert\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.890824 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.891177 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.895377 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.903426 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-72gwc"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.904653 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.909022 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7xhqh"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.912830 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dkss6"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.916441 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.916916 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.922439 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.927567 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.930491 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4z2br"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.931782 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2pqps"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.933084 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.934467 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6sm87"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.935309 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.938547 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.942924 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bbdns"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.944636 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.945265 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5g9hp"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.947571 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.949861 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.951719 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-clh68"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.953963 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.954291 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.955768 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bv2wn"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.959788 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sbgf7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.966392 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mjb8l"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.971783 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.974216 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.974483 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.977523 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.979298 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nvzm5"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.981773 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-whvmw"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.983723 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.985680 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.986120 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rj85t"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.987196 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7kd7r"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.988049 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.989122 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.990137 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xkw2f"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.992198 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.993186 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.994501 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4pxc7"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.995755 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-72gwc"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.996059 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.996717 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.997802 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h6tc4"] Nov 25 10:39:02 crc kubenswrapper[4696]: I1125 10:39:02.998888 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt"] Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.000592 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dkss6"] Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.001456 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.002168 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bbdns"] Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.004028 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4pxc7"] Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.005228 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkn7w"] Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.014344 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.034467 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.055372 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.075741 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.094761 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.115218 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.135725 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.161127 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.174434 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.194946 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.214624 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.235417 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.255656 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.275090 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.294272 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.314754 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.334135 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.354594 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.375099 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.394105 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.414587 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.434604 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.464658 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.478789 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.495336 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.514116 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.535037 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.554955 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.574452 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.595193 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.619629 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.634386 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.654758 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.674438 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.694354 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.713848 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.733797 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.753717 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.773875 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.799858 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.813764 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.836057 4696 request.go:700] Waited for 1.021529821s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.852234 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89c24\" (UniqueName: \"kubernetes.io/projected/735300e4-0a1b-4e3f-9735-84abcf63cf4a-kube-api-access-89c24\") pod \"route-controller-manager-6576b87f9c-4dhfn\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.868399 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxsg7\" (UniqueName: \"kubernetes.io/projected/c830e283-0231-4570-97f7-5583aa7d2c19-kube-api-access-dxsg7\") pod \"openshift-apiserver-operator-796bbdcf4f-t5jvs\" (UID: \"c830e283-0231-4570-97f7-5583aa7d2c19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.883931 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.891691 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdkrm\" (UniqueName: \"kubernetes.io/projected/5939a67f-5765-4afd-905e-76c5a3c8c783-kube-api-access-fdkrm\") pod \"openshift-controller-manager-operator-756b6f6bc6-dmhj7\" (UID: \"5939a67f-5765-4afd-905e-76c5a3c8c783\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.893999 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.929473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m2gd\" (UniqueName: \"kubernetes.io/projected/50d299fa-89e9-41ba-8894-af097502b258-kube-api-access-4m2gd\") pod \"apiserver-7bbb656c7d-mssfq\" (UID: \"50d299fa-89e9-41ba-8894-af097502b258\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.934589 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.951314 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.955395 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.974672 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.989427 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" Nov 25 10:39:03 crc kubenswrapper[4696]: I1125 10:39:03.995285 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.030110 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzntr\" (UniqueName: \"kubernetes.io/projected/0be09268-4279-43e9-8da0-78c4cdd920cd-kube-api-access-mzntr\") pod \"controller-manager-879f6c89f-clh68\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.049773 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvcf\" (UniqueName: \"kubernetes.io/projected/7b36fad9-a6b9-4082-8861-95d4aba0efb7-kube-api-access-4zvcf\") pod \"machine-approver-56656f9798-6b4sd\" (UID: \"7b36fad9-a6b9-4082-8861-95d4aba0efb7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.054919 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.069233 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn"] Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.074067 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 10:39:04 crc kubenswrapper[4696]: W1125 10:39:04.084221 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod735300e4_0a1b_4e3f_9735_84abcf63cf4a.slice/crio-eac984141d5e6b5c91c00b360cba8b01b3588d9dc6c0b5c405ab3b7d62cf058d WatchSource:0}: Error finding container eac984141d5e6b5c91c00b360cba8b01b3588d9dc6c0b5c405ab3b7d62cf058d: Status 404 returned error can't find the container with id eac984141d5e6b5c91c00b360cba8b01b3588d9dc6c0b5c405ab3b7d62cf058d Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.094590 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.114545 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.122629 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7"] Nov 25 10:39:04 crc kubenswrapper[4696]: W1125 10:39:04.128539 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5939a67f_5765_4afd_905e_76c5a3c8c783.slice/crio-6bd77eaeed90d0269cddc2dc1fd7e69db8483a6498c4686a9c64f4c28042ed1f WatchSource:0}: Error finding container 6bd77eaeed90d0269cddc2dc1fd7e69db8483a6498c4686a9c64f4c28042ed1f: Status 404 returned error can't find the container with id 6bd77eaeed90d0269cddc2dc1fd7e69db8483a6498c4686a9c64f4c28042ed1f Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.134780 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.135973 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.172126 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.174230 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.183237 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs"] Nov 25 10:39:04 crc kubenswrapper[4696]: W1125 10:39:04.187167 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b36fad9_a6b9_4082_8861_95d4aba0efb7.slice/crio-e4221b0457648720b0a04115f46da1a637c1800e2215747ff8d57ae323dec756 WatchSource:0}: Error finding container e4221b0457648720b0a04115f46da1a637c1800e2215747ff8d57ae323dec756: Status 404 returned error can't find the container with id e4221b0457648720b0a04115f46da1a637c1800e2215747ff8d57ae323dec756 Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.193801 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: W1125 10:39:04.195854 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc830e283_0231_4570_97f7_5583aa7d2c19.slice/crio-5394306506b16d2b19a5f7f5408fb4a2a32213a3309d47602365bd4c9ff15f7b WatchSource:0}: Error finding container 5394306506b16d2b19a5f7f5408fb4a2a32213a3309d47602365bd4c9ff15f7b: Status 404 returned error can't find the container with id 5394306506b16d2b19a5f7f5408fb4a2a32213a3309d47602365bd4c9ff15f7b Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.214946 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.234936 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.254648 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.264099 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.273916 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.294713 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.313312 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.320820 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq"] Nov 25 10:39:04 crc kubenswrapper[4696]: W1125 10:39:04.327886 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50d299fa_89e9_41ba_8894_af097502b258.slice/crio-d4f04ac440e3ca79103286d1eb5fe887b6aca126c92160dfe2899c631b7c6dcd WatchSource:0}: Error finding container d4f04ac440e3ca79103286d1eb5fe887b6aca126c92160dfe2899c631b7c6dcd: Status 404 returned error can't find the container with id d4f04ac440e3ca79103286d1eb5fe887b6aca126c92160dfe2899c631b7c6dcd Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.342987 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.359359 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.394408 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.414544 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.434875 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438477 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6243043f-a603-4094-9f48-338241d77432-proxy-tls\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438587 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4058138e-66c6-45fe-a684-dc6e7b5b8339-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438626 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-config\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438643 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkbk\" (UniqueName: \"kubernetes.io/projected/84bba043-a192-4ea4-9225-73cc33b31ea5-kube-api-access-sfkbk\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438676 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36c304aa-55a3-49e5-ab60-394509f4b3ca-proxy-tls\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438693 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438764 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krfr4\" (UniqueName: \"kubernetes.io/projected/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-kube-api-access-krfr4\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438782 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0af4515c-aa04-454f-ac8c-412ba49f295a-metrics-tls\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438796 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e9bbce-9822-49de-934c-0ab922daf87e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438893 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36c304aa-55a3-49e5-ab60-394509f4b3ca-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.438929 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nh54\" (UniqueName: \"kubernetes.io/projected/0330c6a6-2aaf-4621-9983-f947a3bacde9-kube-api-access-9nh54\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-metrics-certs\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439168 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5b614b67-512d-474b-ac4e-8cab982cc2f7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6sm87\" (UID: \"5b614b67-512d-474b-ac4e-8cab982cc2f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439203 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439221 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-client\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439245 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xp4n\" (UniqueName: \"kubernetes.io/projected/4058138e-66c6-45fe-a684-dc6e7b5b8339-kube-api-access-7xp4n\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439262 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6243043f-a603-4094-9f48-338241d77432-images\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439276 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0330c6a6-2aaf-4621-9983-f947a3bacde9-config\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439292 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xcz4\" (UniqueName: \"kubernetes.io/projected/a0679b82-8f22-4227-bdbf-d92e063ffba0-kube-api-access-6xcz4\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439322 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84d56ecc-4b70-4776-b906-4228569ddac6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439347 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-service-ca\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439365 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms2zc\" (UniqueName: \"kubernetes.io/projected/36c304aa-55a3-49e5-ab60-394509f4b3ca-kube-api-access-ms2zc\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439396 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0330c6a6-2aaf-4621-9983-f947a3bacde9-serving-cert\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439424 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fv2\" (UniqueName: \"kubernetes.io/projected/23a01a08-d008-451f-86c0-f3c456fafb88-kube-api-access-d8fv2\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439460 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1695d071-bf51-4bed-a09c-5feefd87adfd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wzm5m\" (UID: \"1695d071-bf51-4bed-a09c-5feefd87adfd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439476 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w6zb\" (UniqueName: \"kubernetes.io/projected/e6d2a426-4cba-4b45-bce3-45633b970e8e-kube-api-access-2w6zb\") pod \"downloads-7954f5f757-mjb8l\" (UID: \"e6d2a426-4cba-4b45-bce3-45633b970e8e\") " pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439492 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/23a01a08-d008-451f-86c0-f3c456fafb88-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439513 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgdm7\" (UniqueName: \"kubernetes.io/projected/5b614b67-512d-474b-ac4e-8cab982cc2f7-kube-api-access-fgdm7\") pod \"multus-admission-controller-857f4d67dd-6sm87\" (UID: \"5b614b67-512d-474b-ac4e-8cab982cc2f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439527 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-ca\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-stats-auth\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439585 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps8j2\" (UniqueName: \"kubernetes.io/projected/8a4561e0-205d-4866-8040-c3cec49d3127-kube-api-access-ps8j2\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439612 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0af4515c-aa04-454f-ac8c-412ba49f295a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439627 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e9bbce-9822-49de-934c-0ab922daf87e-config\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439703 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84bba043-a192-4ea4-9225-73cc33b31ea5-service-ca-bundle\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439726 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29w79\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-kube-api-access-29w79\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.439740 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d56ecc-4b70-4776-b906-4228569ddac6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440174 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-bound-sa-token\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6243043f-a603-4094-9f48-338241d77432-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440226 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cpzj\" (UniqueName: \"kubernetes.io/projected/6243043f-a603-4094-9f48-338241d77432-kube-api-access-9cpzj\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440242 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-tls\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440258 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-default-certificate\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440285 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84d56ecc-4b70-4776-b906-4228569ddac6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440301 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a01a08-d008-451f-86c0-f3c456fafb88-config\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440319 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440336 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a4561e0-205d-4866-8040-c3cec49d3127-serving-cert\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.440353 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a0679b82-8f22-4227-bdbf-d92e063ffba0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441397 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-certificates\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441597 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4058138e-66c6-45fe-a684-dc6e7b5b8339-srv-cert\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441632 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/23a01a08-d008-451f-86c0-f3c456fafb88-images\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441647 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0330c6a6-2aaf-4621-9983-f947a3bacde9-trusted-ca\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441744 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67e9bbce-9822-49de-934c-0ab922daf87e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441806 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441830 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0679b82-8f22-4227-bdbf-d92e063ffba0-serving-cert\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441859 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0af4515c-aa04-454f-ac8c-412ba49f295a-trusted-ca\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441878 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ck4c\" (UniqueName: \"kubernetes.io/projected/0af4515c-aa04-454f-ac8c-412ba49f295a-kube-api-access-9ck4c\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441895 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwk6\" (UniqueName: \"kubernetes.io/projected/1695d071-bf51-4bed-a09c-5feefd87adfd-kube-api-access-mgwk6\") pod \"package-server-manager-789f6589d5-wzm5m\" (UID: \"1695d071-bf51-4bed-a09c-5feefd87adfd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.441924 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-trusted-ca\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.442112 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:04.942101722 +0000 UTC m=+141.755718409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.456109 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-clh68"] Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.456241 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.474988 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.495071 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.515424 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.533958 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543261 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.543445 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.043414046 +0000 UTC m=+141.857030633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543494 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6243043f-a603-4094-9f48-338241d77432-images\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543535 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1829a4cf-9696-44fe-a3ec-f3540ac23f81-metrics-tls\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543555 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84d56ecc-4b70-4776-b906-4228569ddac6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543605 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lccb\" (UniqueName: \"kubernetes.io/projected/6ae04990-84f3-4612-9325-0a40352e1278-kube-api-access-8lccb\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543632 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms2zc\" (UniqueName: \"kubernetes.io/projected/36c304aa-55a3-49e5-ab60-394509f4b3ca-kube-api-access-ms2zc\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543714 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.543736 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.544151 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-etcd-serving-ca\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.544175 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjv4\" (UniqueName: \"kubernetes.io/projected/106060a8-1b39-40b0-ac39-b0f057c9552c-kube-api-access-jmjv4\") pod \"cluster-samples-operator-665b6dd947-zk5fp\" (UID: \"106060a8-1b39-40b0-ac39-b0f057c9552c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.544238 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1695d071-bf51-4bed-a09c-5feefd87adfd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wzm5m\" (UID: \"1695d071-bf51-4bed-a09c-5feefd87adfd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.544257 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w6zb\" (UniqueName: \"kubernetes.io/projected/e6d2a426-4cba-4b45-bce3-45633b970e8e-kube-api-access-2w6zb\") pod \"downloads-7954f5f757-mjb8l\" (UID: \"e6d2a426-4cba-4b45-bce3-45633b970e8e\") " pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.544866 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6243043f-a603-4094-9f48-338241d77432-images\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545210 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545262 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51aacc61-4839-41db-a5ae-a92738581de1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545307 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgdm7\" (UniqueName: \"kubernetes.io/projected/5b614b67-512d-474b-ac4e-8cab982cc2f7-kube-api-access-fgdm7\") pod \"multus-admission-controller-857f4d67dd-6sm87\" (UID: \"5b614b67-512d-474b-ac4e-8cab982cc2f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545332 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-serving-cert\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545498 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/29a2d684-8d98-42c3-9d91-12d0a87bf032-audit-dir\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545611 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-serving-cert\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545722 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-stats-auth\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545743 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-oauth-config\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545770 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/162ad055-0833-461f-92e6-496d1bb33e21-tmpfs\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545790 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-etcd-client\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545813 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0af4515c-aa04-454f-ac8c-412ba49f295a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545837 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e9bbce-9822-49de-934c-0ab922daf87e-config\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545864 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s8gz\" (UniqueName: \"kubernetes.io/projected/51aacc61-4839-41db-a5ae-a92738581de1-kube-api-access-5s8gz\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545884 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d56ecc-4b70-4776-b906-4228569ddac6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545907 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29w79\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-kube-api-access-29w79\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545949 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.545989 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-tls\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546010 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-oauth-serving-cert\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546037 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546061 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a4561e0-205d-4866-8040-c3cec49d3127-serving-cert\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546087 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a0679b82-8f22-4227-bdbf-d92e063ffba0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546138 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0330c6a6-2aaf-4621-9983-f947a3bacde9-trusted-ca\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546164 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67e9bbce-9822-49de-934c-0ab922daf87e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546190 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-dir\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546222 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546247 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0679b82-8f22-4227-bdbf-d92e063ffba0-serving-cert\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546270 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-service-ca\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546298 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwk6\" (UniqueName: \"kubernetes.io/projected/1695d071-bf51-4bed-a09c-5feefd87adfd-kube-api-access-mgwk6\") pod \"package-server-manager-789f6589d5-wzm5m\" (UID: \"1695d071-bf51-4bed-a09c-5feefd87adfd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546323 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546345 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-config\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546368 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt5g6\" (UniqueName: \"kubernetes.io/projected/e4449b1d-871a-4134-bc0a-0cbf2508e10b-kube-api-access-rt5g6\") pod \"dns-operator-744455d44c-whvmw\" (UID: \"e4449b1d-871a-4134-bc0a-0cbf2508e10b\") " pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546389 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-trusted-ca\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546408 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc6b507a-226b-42ab-9a86-d4376ba9dfd7-cert\") pod \"ingress-canary-4pxc7\" (UID: \"dc6b507a-226b-42ab-9a86-d4376ba9dfd7\") " pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546430 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-serving-cert\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546451 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-audit\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546469 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zddfl\" (UniqueName: \"kubernetes.io/projected/29a2d684-8d98-42c3-9d91-12d0a87bf032-kube-api-access-zddfl\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546489 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546509 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-socket-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546534 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4058138e-66c6-45fe-a684-dc6e7b5b8339-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546557 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546579 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-policies\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546604 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6ae04990-84f3-4612-9325-0a40352e1278-signing-cabundle\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546625 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24xbk\" (UniqueName: \"kubernetes.io/projected/f8aa7eb9-5023-45b9-b7b6-10d4feda652c-kube-api-access-24xbk\") pod \"migrator-59844c95c7-vt4qt\" (UID: \"f8aa7eb9-5023-45b9-b7b6-10d4feda652c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546685 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7vh9\" (UniqueName: \"kubernetes.io/projected/376ea9c4-9e68-4475-b205-95e2185e51ba-kube-api-access-w7vh9\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546708 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-config\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546734 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-metrics-certs\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546753 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546774 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5b614b67-512d-474b-ac4e-8cab982cc2f7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6sm87\" (UID: \"5b614b67-512d-474b-ac4e-8cab982cc2f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546791 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jll8\" (UniqueName: \"kubernetes.io/projected/0e6edd3c-d469-4a55-a9b8-2a3087248db6-kube-api-access-7jll8\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546806 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/29a2d684-8d98-42c3-9d91-12d0a87bf032-node-pullsecrets\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546822 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-client\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546835 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1829a4cf-9696-44fe-a3ec-f3540ac23f81-config-volume\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546849 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e6edd3c-d469-4a55-a9b8-2a3087248db6-config-volume\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546865 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xp4n\" (UniqueName: \"kubernetes.io/projected/4058138e-66c6-45fe-a684-dc6e7b5b8339-kube-api-access-7xp4n\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546879 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0330c6a6-2aaf-4621-9983-f947a3bacde9-config\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546894 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xcz4\" (UniqueName: \"kubernetes.io/projected/a0679b82-8f22-4227-bdbf-d92e063ffba0-kube-api-access-6xcz4\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546909 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-console-config\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546925 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-service-ca\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546939 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0330c6a6-2aaf-4621-9983-f947a3bacde9-serving-cert\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546954 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546977 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fv2\" (UniqueName: \"kubernetes.io/projected/23a01a08-d008-451f-86c0-f3c456fafb88-kube-api-access-d8fv2\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.546993 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-config\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547008 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ctk8\" (UniqueName: \"kubernetes.io/projected/dc6b507a-226b-42ab-9a86-d4376ba9dfd7-kube-api-access-2ctk8\") pod \"ingress-canary-4pxc7\" (UID: \"dc6b507a-226b-42ab-9a86-d4376ba9dfd7\") " pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547023 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvfx\" (UniqueName: \"kubernetes.io/projected/162ad055-0833-461f-92e6-496d1bb33e21-kube-api-access-bhvfx\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547044 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/23a01a08-d008-451f-86c0-f3c456fafb88-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547059 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/162ad055-0833-461f-92e6-496d1bb33e21-webhook-cert\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547073 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-mountpoint-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547090 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-ca\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547104 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547122 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c715c60a-acd3-45e9-81df-98802b303a48-profile-collector-cert\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547140 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51aacc61-4839-41db-a5ae-a92738581de1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547164 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps8j2\" (UniqueName: \"kubernetes.io/projected/8a4561e0-205d-4866-8040-c3cec49d3127-kube-api-access-ps8j2\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547183 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547203 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-registration-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547219 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547633 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdmth\" (UniqueName: \"kubernetes.io/projected/c715c60a-acd3-45e9-81df-98802b303a48-kube-api-access-mdmth\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547682 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84bba043-a192-4ea4-9225-73cc33b31ea5-service-ca-bundle\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547717 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e6edd3c-d469-4a55-a9b8-2a3087248db6-secret-volume\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547752 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6243043f-a603-4094-9f48-338241d77432-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547769 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cpzj\" (UniqueName: \"kubernetes.io/projected/6243043f-a603-4094-9f48-338241d77432-kube-api-access-9cpzj\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547789 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-bound-sa-token\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547806 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c715c60a-acd3-45e9-81df-98802b303a48-srv-cert\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547820 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/90f9987e-5e7b-40a3-a750-dac2b017942a-certs\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547835 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-encryption-config\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547852 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-default-certificate\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547867 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-trusted-ca-bundle\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547883 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qhfz\" (UniqueName: \"kubernetes.io/projected/78279129-b25e-4e7c-b563-6043c9ff4f88-kube-api-access-2qhfz\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547927 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84d56ecc-4b70-4776-b906-4228569ddac6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547943 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a01a08-d008-451f-86c0-f3c456fafb88-config\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547960 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc9lw\" (UniqueName: \"kubernetes.io/projected/1829a4cf-9696-44fe-a3ec-f3540ac23f81-kube-api-access-tc9lw\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547974 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5jg6\" (UniqueName: \"kubernetes.io/projected/020fd783-0219-4f9d-8504-f430877d9ed4-kube-api-access-b5jg6\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.547991 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9r5l\" (UniqueName: \"kubernetes.io/projected/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-kube-api-access-f9r5l\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548009 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548026 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548043 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548063 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-certificates\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548081 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4058138e-66c6-45fe-a684-dc6e7b5b8339-srv-cert\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548098 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/23a01a08-d008-451f-86c0-f3c456fafb88-images\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548112 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-plugins-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4449b1d-871a-4134-bc0a-0cbf2508e10b-metrics-tls\") pod \"dns-operator-744455d44c-whvmw\" (UID: \"e4449b1d-871a-4134-bc0a-0cbf2508e10b\") " pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548154 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548175 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0af4515c-aa04-454f-ac8c-412ba49f295a-trusted-ca\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548192 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ck4c\" (UniqueName: \"kubernetes.io/projected/0af4515c-aa04-454f-ac8c-412ba49f295a-kube-api-access-9ck4c\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548208 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86ks8\" (UniqueName: \"kubernetes.io/projected/19c82916-91f4-4a0f-9beb-a65cdea01a8a-kube-api-access-86ks8\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548224 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gmqm\" (UniqueName: \"kubernetes.io/projected/90f9987e-5e7b-40a3-a750-dac2b017942a-kube-api-access-6gmqm\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548240 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548258 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/162ad055-0833-461f-92e6-496d1bb33e21-apiservice-cert\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548278 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d56ecc-4b70-4776-b906-4228569ddac6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548324 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6243043f-a603-4094-9f48-338241d77432-proxy-tls\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548343 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/106060a8-1b39-40b0-ac39-b0f057c9552c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zk5fp\" (UID: \"106060a8-1b39-40b0-ac39-b0f057c9552c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548360 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkbk\" (UniqueName: \"kubernetes.io/projected/84bba043-a192-4ea4-9225-73cc33b31ea5-kube-api-access-sfkbk\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548374 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-image-import-ca\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548390 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/51aacc61-4839-41db-a5ae-a92738581de1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-config\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548457 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0af4515c-aa04-454f-ac8c-412ba49f295a-metrics-tls\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548473 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e9bbce-9822-49de-934c-0ab922daf87e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.548490 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36c304aa-55a3-49e5-ab60-394509f4b3ca-proxy-tls\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.551124 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krfr4\" (UniqueName: \"kubernetes.io/projected/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-kube-api-access-krfr4\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.551204 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8jsw\" (UniqueName: \"kubernetes.io/projected/2e1aecfe-07a2-4772-bab2-52b5f5fed397-kube-api-access-q8jsw\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.551429 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a0679b82-8f22-4227-bdbf-d92e063ffba0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.551704 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1695d071-bf51-4bed-a09c-5feefd87adfd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wzm5m\" (UID: \"1695d071-bf51-4bed-a09c-5feefd87adfd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.551911 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84d56ecc-4b70-4776-b906-4228569ddac6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.552099 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e9bbce-9822-49de-934c-0ab922daf87e-config\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.552431 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-metrics-certs\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.552804 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.552830 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.052815455 +0000 UTC m=+141.866432032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.553019 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0330c6a6-2aaf-4621-9983-f947a3bacde9-config\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.553881 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.553884 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-service-ca\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.554403 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a4561e0-205d-4866-8040-c3cec49d3127-serving-cert\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.554802 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-tls\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.558333 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0330c6a6-2aaf-4621-9983-f947a3bacde9-trusted-ca\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.560492 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-trusted-ca\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.560973 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6243043f-a603-4094-9f48-338241d77432-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.561690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-ca\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.562754 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84bba043-a192-4ea4-9225-73cc33b31ea5-service-ca-bundle\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.563012 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a01a08-d008-451f-86c0-f3c456fafb88-config\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.564710 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4561e0-205d-4866-8040-c3cec49d3127-config\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.565839 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566090 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c82916-91f4-4a0f-9beb-a65cdea01a8a-config\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566160 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f6f8fdc-fec0-46b9-9277-2996da698f24-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-n62qw\" (UID: \"2f6f8fdc-fec0-46b9-9277-2996da698f24\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566179 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566205 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36c304aa-55a3-49e5-ab60-394509f4b3ca-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566228 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nh54\" (UniqueName: \"kubernetes.io/projected/0330c6a6-2aaf-4621-9983-f947a3bacde9-kube-api-access-9nh54\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566302 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19c82916-91f4-4a0f-9beb-a65cdea01a8a-serving-cert\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566325 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566347 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn28w\" (UniqueName: \"kubernetes.io/projected/2f6f8fdc-fec0-46b9-9277-2996da698f24-kube-api-access-dn28w\") pod \"control-plane-machine-set-operator-78cbb6b69f-n62qw\" (UID: \"2f6f8fdc-fec0-46b9-9277-2996da698f24\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566375 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-csi-data-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566404 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-service-ca-bundle\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566430 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/90f9987e-5e7b-40a3-a750-dac2b017942a-node-bootstrap-token\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566451 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6ae04990-84f3-4612-9325-0a40352e1278-signing-key\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566473 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.566518 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/23a01a08-d008-451f-86c0-f3c456fafb88-images\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.569198 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0af4515c-aa04-454f-ac8c-412ba49f295a-trusted-ca\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.570317 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36c304aa-55a3-49e5-ab60-394509f4b3ca-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.570482 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-certificates\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.571620 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.572099 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8a4561e0-205d-4866-8040-c3cec49d3127-etcd-client\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.572116 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e9bbce-9822-49de-934c-0ab922daf87e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.572471 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-stats-auth\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.573966 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5b614b67-512d-474b-ac4e-8cab982cc2f7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6sm87\" (UID: \"5b614b67-512d-474b-ac4e-8cab982cc2f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.574516 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.575204 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0679b82-8f22-4227-bdbf-d92e063ffba0-serving-cert\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.575241 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/84bba043-a192-4ea4-9225-73cc33b31ea5-default-certificate\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.575541 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6243043f-a603-4094-9f48-338241d77432-proxy-tls\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.575726 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/23a01a08-d008-451f-86c0-f3c456fafb88-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.575952 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36c304aa-55a3-49e5-ab60-394509f4b3ca-proxy-tls\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.576062 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0330c6a6-2aaf-4621-9983-f947a3bacde9-serving-cert\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.576407 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4058138e-66c6-45fe-a684-dc6e7b5b8339-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.577553 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4058138e-66c6-45fe-a684-dc6e7b5b8339-srv-cert\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.577555 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0af4515c-aa04-454f-ac8c-412ba49f295a-metrics-tls\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.595045 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.614720 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.634911 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.654213 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.667753 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.667991 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/106060a8-1b39-40b0-ac39-b0f057c9552c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zk5fp\" (UID: \"106060a8-1b39-40b0-ac39-b0f057c9552c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668026 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/162ad055-0833-461f-92e6-496d1bb33e21-apiservice-cert\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668088 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-image-import-ca\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668111 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/51aacc61-4839-41db-a5ae-a92738581de1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668138 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c82916-91f4-4a0f-9beb-a65cdea01a8a-config\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668156 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668176 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8jsw\" (UniqueName: \"kubernetes.io/projected/2e1aecfe-07a2-4772-bab2-52b5f5fed397-kube-api-access-q8jsw\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668217 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f6f8fdc-fec0-46b9-9277-2996da698f24-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-n62qw\" (UID: \"2f6f8fdc-fec0-46b9-9277-2996da698f24\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668241 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19c82916-91f4-4a0f-9beb-a65cdea01a8a-serving-cert\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668263 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668313 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn28w\" (UniqueName: \"kubernetes.io/projected/2f6f8fdc-fec0-46b9-9277-2996da698f24-kube-api-access-dn28w\") pod \"control-plane-machine-set-operator-78cbb6b69f-n62qw\" (UID: \"2f6f8fdc-fec0-46b9-9277-2996da698f24\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-csi-data-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668355 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/90f9987e-5e7b-40a3-a750-dac2b017942a-node-bootstrap-token\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668374 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6ae04990-84f3-4612-9325-0a40352e1278-signing-key\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668394 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-service-ca-bundle\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668417 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1829a4cf-9696-44fe-a3ec-f3540ac23f81-metrics-tls\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668440 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lccb\" (UniqueName: \"kubernetes.io/projected/6ae04990-84f3-4612-9325-0a40352e1278-kube-api-access-8lccb\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668473 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668494 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-etcd-serving-ca\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668522 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmjv4\" (UniqueName: \"kubernetes.io/projected/106060a8-1b39-40b0-ac39-b0f057c9552c-kube-api-access-jmjv4\") pod \"cluster-samples-operator-665b6dd947-zk5fp\" (UID: \"106060a8-1b39-40b0-ac39-b0f057c9552c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668553 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-serving-cert\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668574 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51aacc61-4839-41db-a5ae-a92738581de1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668595 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-oauth-config\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668616 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/29a2d684-8d98-42c3-9d91-12d0a87bf032-audit-dir\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668637 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-serving-cert\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668729 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/162ad055-0833-461f-92e6-496d1bb33e21-tmpfs\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668751 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-etcd-client\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668771 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s8gz\" (UniqueName: \"kubernetes.io/projected/51aacc61-4839-41db-a5ae-a92738581de1-kube-api-access-5s8gz\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668798 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-oauth-serving-cert\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668818 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668840 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668871 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-service-ca\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668890 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-dir\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668921 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-config\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668941 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt5g6\" (UniqueName: \"kubernetes.io/projected/e4449b1d-871a-4134-bc0a-0cbf2508e10b-kube-api-access-rt5g6\") pod \"dns-operator-744455d44c-whvmw\" (UID: \"e4449b1d-871a-4134-bc0a-0cbf2508e10b\") " pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668966 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.668988 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc6b507a-226b-42ab-9a86-d4376ba9dfd7-cert\") pod \"ingress-canary-4pxc7\" (UID: \"dc6b507a-226b-42ab-9a86-d4376ba9dfd7\") " pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669023 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-serving-cert\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669046 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zddfl\" (UniqueName: \"kubernetes.io/projected/29a2d684-8d98-42c3-9d91-12d0a87bf032-kube-api-access-zddfl\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669067 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-audit\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669084 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669102 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-socket-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669120 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-policies\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669140 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6ae04990-84f3-4612-9325-0a40352e1278-signing-cabundle\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669158 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7vh9\" (UniqueName: \"kubernetes.io/projected/376ea9c4-9e68-4475-b205-95e2185e51ba-kube-api-access-w7vh9\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-config\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669197 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24xbk\" (UniqueName: \"kubernetes.io/projected/f8aa7eb9-5023-45b9-b7b6-10d4feda652c-kube-api-access-24xbk\") pod \"migrator-59844c95c7-vt4qt\" (UID: \"f8aa7eb9-5023-45b9-b7b6-10d4feda652c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669220 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669244 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1829a4cf-9696-44fe-a3ec-f3540ac23f81-config-volume\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669264 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e6edd3c-d469-4a55-a9b8-2a3087248db6-config-volume\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669285 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jll8\" (UniqueName: \"kubernetes.io/projected/0e6edd3c-d469-4a55-a9b8-2a3087248db6-kube-api-access-7jll8\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669310 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/29a2d684-8d98-42c3-9d91-12d0a87bf032-node-pullsecrets\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669348 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-console-config\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669371 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669394 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-config\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669435 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ctk8\" (UniqueName: \"kubernetes.io/projected/dc6b507a-226b-42ab-9a86-d4376ba9dfd7-kube-api-access-2ctk8\") pod \"ingress-canary-4pxc7\" (UID: \"dc6b507a-226b-42ab-9a86-d4376ba9dfd7\") " pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669459 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvfx\" (UniqueName: \"kubernetes.io/projected/162ad055-0833-461f-92e6-496d1bb33e21-kube-api-access-bhvfx\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669481 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669503 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/162ad055-0833-461f-92e6-496d1bb33e21-webhook-cert\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669524 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-mountpoint-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669546 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c715c60a-acd3-45e9-81df-98802b303a48-profile-collector-cert\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669572 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-registration-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669613 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51aacc61-4839-41db-a5ae-a92738581de1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669639 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdmth\" (UniqueName: \"kubernetes.io/projected/c715c60a-acd3-45e9-81df-98802b303a48-kube-api-access-mdmth\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669685 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669725 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e6edd3c-d469-4a55-a9b8-2a3087248db6-secret-volume\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669761 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-trusted-ca-bundle\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669783 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qhfz\" (UniqueName: \"kubernetes.io/projected/78279129-b25e-4e7c-b563-6043c9ff4f88-kube-api-access-2qhfz\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669801 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c715c60a-acd3-45e9-81df-98802b303a48-srv-cert\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669821 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/90f9987e-5e7b-40a3-a750-dac2b017942a-certs\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669837 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-encryption-config\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669853 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc9lw\" (UniqueName: \"kubernetes.io/projected/1829a4cf-9696-44fe-a3ec-f3540ac23f81-kube-api-access-tc9lw\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669867 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5jg6\" (UniqueName: \"kubernetes.io/projected/020fd783-0219-4f9d-8504-f430877d9ed4-kube-api-access-b5jg6\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669889 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669905 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9r5l\" (UniqueName: \"kubernetes.io/projected/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-kube-api-access-f9r5l\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669922 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669937 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4449b1d-871a-4134-bc0a-0cbf2508e10b-metrics-tls\") pod \"dns-operator-744455d44c-whvmw\" (UID: \"e4449b1d-871a-4134-bc0a-0cbf2508e10b\") " pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669951 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-plugins-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669971 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86ks8\" (UniqueName: \"kubernetes.io/projected/19c82916-91f4-4a0f-9beb-a65cdea01a8a-kube-api-access-86ks8\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.669988 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.670013 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.670034 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gmqm\" (UniqueName: \"kubernetes.io/projected/90f9987e-5e7b-40a3-a750-dac2b017942a-kube-api-access-6gmqm\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.670819 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-audit\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.671160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-socket-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.671227 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.671323 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.171305318 +0000 UTC m=+141.984921905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.671858 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-registration-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.672148 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-policies\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.672478 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-etcd-serving-ca\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.672780 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6ae04990-84f3-4612-9325-0a40352e1278-signing-cabundle\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.673152 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-serving-cert\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.673282 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51aacc61-4839-41db-a5ae-a92738581de1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.673474 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-config\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.673815 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-oauth-serving-cert\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.674273 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.675158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1829a4cf-9696-44fe-a3ec-f3540ac23f81-config-volume\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.675278 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.675300 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-csi-data-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.675943 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e6edd3c-d469-4a55-a9b8-2a3087248db6-config-volume\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.676044 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/29a2d684-8d98-42c3-9d91-12d0a87bf032-node-pullsecrets\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.676586 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-console-config\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.676798 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/106060a8-1b39-40b0-ac39-b0f057c9552c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zk5fp\" (UID: \"106060a8-1b39-40b0-ac39-b0f057c9552c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.676832 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.676876 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-dir\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.676881 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-service-ca-bundle\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.677587 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-config\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.678564 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6ae04990-84f3-4612-9325-0a40352e1278-signing-key\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.678974 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.679153 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-service-ca\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.679469 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.679533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/90f9987e-5e7b-40a3-a750-dac2b017942a-node-bootstrap-token\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.679635 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-serving-cert\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.680064 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/29a2d684-8d98-42c3-9d91-12d0a87bf032-audit-dir\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.680138 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-plugins-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.680214 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.680239 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-config\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.680790 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681156 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681536 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681606 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e6edd3c-d469-4a55-a9b8-2a3087248db6-secret-volume\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681707 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681733 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/020fd783-0219-4f9d-8504-f430877d9ed4-mountpoint-dir\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681779 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-trusted-ca-bundle\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.681890 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.682291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-oauth-config\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.684745 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.684983 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c82916-91f4-4a0f-9beb-a65cdea01a8a-config\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.685064 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-serving-cert\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.685254 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c715c60a-acd3-45e9-81df-98802b303a48-srv-cert\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.685316 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.685686 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/162ad055-0833-461f-92e6-496d1bb33e21-tmpfs\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.685987 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4449b1d-871a-4134-bc0a-0cbf2508e10b-metrics-tls\") pod \"dns-operator-744455d44c-whvmw\" (UID: \"e4449b1d-871a-4134-bc0a-0cbf2508e10b\") " pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.686043 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/162ad055-0833-461f-92e6-496d1bb33e21-webhook-cert\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.686110 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.686439 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f6f8fdc-fec0-46b9-9277-2996da698f24-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-n62qw\" (UID: \"2f6f8fdc-fec0-46b9-9277-2996da698f24\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.686471 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19c82916-91f4-4a0f-9beb-a65cdea01a8a-serving-cert\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.686602 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29a2d684-8d98-42c3-9d91-12d0a87bf032-image-import-ca\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.687035 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.687158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/51aacc61-4839-41db-a5ae-a92738581de1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.687813 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.687922 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/162ad055-0833-461f-92e6-496d1bb33e21-apiservice-cert\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.688188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c715c60a-acd3-45e9-81df-98802b303a48-profile-collector-cert\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.688487 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-etcd-client\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.689299 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/90f9987e-5e7b-40a3-a750-dac2b017942a-certs\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.690264 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29a2d684-8d98-42c3-9d91-12d0a87bf032-encryption-config\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.693864 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.701197 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1829a4cf-9696-44fe-a3ec-f3540ac23f81-metrics-tls\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.714291 4696 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.734444 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.753940 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.770940 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.771270 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.271255709 +0000 UTC m=+142.084872296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.773764 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.794016 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.811535 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" event={"ID":"c830e283-0231-4570-97f7-5583aa7d2c19","Type":"ContainerStarted","Data":"c4e9884ce3168d3bf1585e68fdd956d61f97376d9ee758aca56952e612edd612"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.811628 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" event={"ID":"c830e283-0231-4570-97f7-5583aa7d2c19","Type":"ContainerStarted","Data":"5394306506b16d2b19a5f7f5408fb4a2a32213a3309d47602365bd4c9ff15f7b"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.814854 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.815274 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" event={"ID":"735300e4-0a1b-4e3f-9735-84abcf63cf4a","Type":"ContainerStarted","Data":"ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.815327 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" event={"ID":"735300e4-0a1b-4e3f-9735-84abcf63cf4a","Type":"ContainerStarted","Data":"eac984141d5e6b5c91c00b360cba8b01b3588d9dc6c0b5c405ab3b7d62cf058d"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.816201 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.818191 4696 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4dhfn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.818241 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" podUID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.819656 4696 generic.go:334] "Generic (PLEG): container finished" podID="50d299fa-89e9-41ba-8894-af097502b258" containerID="f89d5baebb9fa1f4f9ccdacea25d869459570cfb8220cd362e6c9c7f9fd527e0" exitCode=0 Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.819725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" event={"ID":"50d299fa-89e9-41ba-8894-af097502b258","Type":"ContainerDied","Data":"f89d5baebb9fa1f4f9ccdacea25d869459570cfb8220cd362e6c9c7f9fd527e0"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.819749 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" event={"ID":"50d299fa-89e9-41ba-8894-af097502b258","Type":"ContainerStarted","Data":"d4f04ac440e3ca79103286d1eb5fe887b6aca126c92160dfe2899c631b7c6dcd"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.822654 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" event={"ID":"0be09268-4279-43e9-8da0-78c4cdd920cd","Type":"ContainerStarted","Data":"e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.822707 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" event={"ID":"0be09268-4279-43e9-8da0-78c4cdd920cd","Type":"ContainerStarted","Data":"faff271f3ba07160ea5a28d9a5338c4c34b1d70b15548cce624a6508945648a0"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.825437 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" event={"ID":"7b36fad9-a6b9-4082-8861-95d4aba0efb7","Type":"ContainerStarted","Data":"08bc31de087a596f35fbb127e632828679be3db4e078e2bb5868bbd4c3e1a9c7"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.825485 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" event={"ID":"7b36fad9-a6b9-4082-8861-95d4aba0efb7","Type":"ContainerStarted","Data":"8048bc9424e8525d58606b004a4777533b1f328a6a28095b346dd54502f3fe52"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.825495 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" event={"ID":"7b36fad9-a6b9-4082-8861-95d4aba0efb7","Type":"ContainerStarted","Data":"e4221b0457648720b0a04115f46da1a637c1800e2215747ff8d57ae323dec756"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.827636 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" event={"ID":"5939a67f-5765-4afd-905e-76c5a3c8c783","Type":"ContainerStarted","Data":"1c877e9464c818f7380fb239c205c60289e2f0e723b301cf23ccd0ba7776a401"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.827687 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" event={"ID":"5939a67f-5765-4afd-905e-76c5a3c8c783","Type":"ContainerStarted","Data":"6bd77eaeed90d0269cddc2dc1fd7e69db8483a6498c4686a9c64f4c28042ed1f"} Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.834492 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.842188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc6b507a-226b-42ab-9a86-d4376ba9dfd7-cert\") pod \"ingress-canary-4pxc7\" (UID: \"dc6b507a-226b-42ab-9a86-d4376ba9dfd7\") " pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.872151 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.872421 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.372400817 +0000 UTC m=+142.186017424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.872859 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.873180 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.373170311 +0000 UTC m=+142.186786898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.892363 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms2zc\" (UniqueName: \"kubernetes.io/projected/36c304aa-55a3-49e5-ab60-394509f4b3ca-kube-api-access-ms2zc\") pod \"machine-config-controller-84d6567774-hkt6s\" (UID: \"36c304aa-55a3-49e5-ab60-394509f4b3ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.909735 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w6zb\" (UniqueName: \"kubernetes.io/projected/e6d2a426-4cba-4b45-bce3-45633b970e8e-kube-api-access-2w6zb\") pod \"downloads-7954f5f757-mjb8l\" (UID: \"e6d2a426-4cba-4b45-bce3-45633b970e8e\") " pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.928608 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgdm7\" (UniqueName: \"kubernetes.io/projected/5b614b67-512d-474b-ac4e-8cab982cc2f7-kube-api-access-fgdm7\") pod \"multus-admission-controller-857f4d67dd-6sm87\" (UID: \"5b614b67-512d-474b-ac4e-8cab982cc2f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.951525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67e9bbce-9822-49de-934c-0ab922daf87e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-s4kcg\" (UID: \"67e9bbce-9822-49de-934c-0ab922daf87e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.967971 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29w79\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-kube-api-access-29w79\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.973761 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.973879 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.473862835 +0000 UTC m=+142.287479422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.977325 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:04 crc kubenswrapper[4696]: E1125 10:39:04.978579 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.4785637 +0000 UTC m=+142.292180287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.988930 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0af4515c-aa04-454f-ac8c-412ba49f295a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:04 crc kubenswrapper[4696]: I1125 10:39:04.990853 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.017956 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xcz4\" (UniqueName: \"kubernetes.io/projected/a0679b82-8f22-4227-bdbf-d92e063ffba0-kube-api-access-6xcz4\") pod \"openshift-config-operator-7777fb866f-2pqps\" (UID: \"a0679b82-8f22-4227-bdbf-d92e063ffba0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.033224 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fv2\" (UniqueName: \"kubernetes.io/projected/23a01a08-d008-451f-86c0-f3c456fafb88-kube-api-access-d8fv2\") pod \"machine-api-operator-5694c8668f-5g9hp\" (UID: \"23a01a08-d008-451f-86c0-f3c456fafb88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.047417 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.049606 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwk6\" (UniqueName: \"kubernetes.io/projected/1695d071-bf51-4bed-a09c-5feefd87adfd-kube-api-access-mgwk6\") pod \"package-server-manager-789f6589d5-wzm5m\" (UID: \"1695d071-bf51-4bed-a09c-5feefd87adfd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.057939 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.076350 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps8j2\" (UniqueName: \"kubernetes.io/projected/8a4561e0-205d-4866-8040-c3cec49d3127-kube-api-access-ps8j2\") pod \"etcd-operator-b45778765-4z2br\" (UID: \"8a4561e0-205d-4866-8040-c3cec49d3127\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.079599 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.079776 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.579751199 +0000 UTC m=+142.393367786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.080102 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.080878 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.580856204 +0000 UTC m=+142.394472791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.089485 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cpzj\" (UniqueName: \"kubernetes.io/projected/6243043f-a603-4094-9f48-338241d77432-kube-api-access-9cpzj\") pod \"machine-config-operator-74547568cd-6wkz7\" (UID: \"6243043f-a603-4094-9f48-338241d77432\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.089906 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.113852 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.119461 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84d56ecc-4b70-4776-b906-4228569ddac6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6dbht\" (UID: \"84d56ecc-4b70-4776-b906-4228569ddac6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.135208 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-bound-sa-token\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.150608 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xp4n\" (UniqueName: \"kubernetes.io/projected/4058138e-66c6-45fe-a684-dc6e7b5b8339-kube-api-access-7xp4n\") pod \"olm-operator-6b444d44fb-qwr48\" (UID: \"4058138e-66c6-45fe-a684-dc6e7b5b8339\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.177187 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ck4c\" (UniqueName: \"kubernetes.io/projected/0af4515c-aa04-454f-ac8c-412ba49f295a-kube-api-access-9ck4c\") pod \"ingress-operator-5b745b69d9-b82sf\" (UID: \"0af4515c-aa04-454f-ac8c-412ba49f295a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.186689 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.187444 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.687421369 +0000 UTC m=+142.501037956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.190754 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.195555 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkbk\" (UniqueName: \"kubernetes.io/projected/84bba043-a192-4ea4-9225-73cc33b31ea5-kube-api-access-sfkbk\") pod \"router-default-5444994796-k24kz\" (UID: \"84bba043-a192-4ea4-9225-73cc33b31ea5\") " pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.237283 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nh54\" (UniqueName: \"kubernetes.io/projected/0330c6a6-2aaf-4621-9983-f947a3bacde9-kube-api-access-9nh54\") pod \"console-operator-58897d9998-sbgf7\" (UID: \"0330c6a6-2aaf-4621-9983-f947a3bacde9\") " pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.247942 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krfr4\" (UniqueName: \"kubernetes.io/projected/e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9-kube-api-access-krfr4\") pod \"kube-storage-version-migrator-operator-b67b599dd-4sg27\" (UID: \"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.254401 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gmqm\" (UniqueName: \"kubernetes.io/projected/90f9987e-5e7b-40a3-a750-dac2b017942a-kube-api-access-6gmqm\") pod \"machine-config-server-7xhqh\" (UID: \"90f9987e-5e7b-40a3-a750-dac2b017942a\") " pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.273321 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zddfl\" (UniqueName: \"kubernetes.io/projected/29a2d684-8d98-42c3-9d91-12d0a87bf032-kube-api-access-zddfl\") pod \"apiserver-76f77b778f-bv2wn\" (UID: \"29a2d684-8d98-42c3-9d91-12d0a87bf032\") " pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.277378 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.298817 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.299339 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.799325029 +0000 UTC m=+142.612941616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.300252 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6sm87"] Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.312133 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.316340 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s8gz\" (UniqueName: \"kubernetes.io/projected/51aacc61-4839-41db-a5ae-a92738581de1-kube-api-access-5s8gz\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.318838 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbc2f0e2-9d49-4d3c-9c44-15b762505b83-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-g5n86\" (UID: \"bbc2f0e2-9d49-4d3c-9c44-15b762505b83\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.332105 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.334183 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.344211 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.368134 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.376610 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7vh9\" (UniqueName: \"kubernetes.io/projected/376ea9c4-9e68-4475-b205-95e2185e51ba-kube-api-access-w7vh9\") pod \"console-f9d7485db-rj85t\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.387469 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.392822 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdmth\" (UniqueName: \"kubernetes.io/projected/c715c60a-acd3-45e9-81df-98802b303a48-kube-api-access-mdmth\") pod \"catalog-operator-68c6474976-lpgdr\" (UID: \"c715c60a-acd3-45e9-81df-98802b303a48\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.397291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmjv4\" (UniqueName: \"kubernetes.io/projected/106060a8-1b39-40b0-ac39-b0f057c9552c-kube-api-access-jmjv4\") pod \"cluster-samples-operator-665b6dd947-zk5fp\" (UID: \"106060a8-1b39-40b0-ac39-b0f057c9552c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.403540 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.405310 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.405606 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.405780 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.905741239 +0000 UTC m=+142.719357826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.406082 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.406781 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:05.90676365 +0000 UTC m=+142.720380237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.412249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24xbk\" (UniqueName: \"kubernetes.io/projected/f8aa7eb9-5023-45b9-b7b6-10d4feda652c-kube-api-access-24xbk\") pod \"migrator-59844c95c7-vt4qt\" (UID: \"f8aa7eb9-5023-45b9-b7b6-10d4feda652c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.434632 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.436093 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.440238 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5jg6\" (UniqueName: \"kubernetes.io/projected/020fd783-0219-4f9d-8504-f430877d9ed4-kube-api-access-b5jg6\") pod \"csi-hostpathplugin-bbdns\" (UID: \"020fd783-0219-4f9d-8504-f430877d9ed4\") " pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.445778 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.452129 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn28w\" (UniqueName: \"kubernetes.io/projected/2f6f8fdc-fec0-46b9-9277-2996da698f24-kube-api-access-dn28w\") pod \"control-plane-machine-set-operator-78cbb6b69f-n62qw\" (UID: \"2f6f8fdc-fec0-46b9-9277-2996da698f24\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.467253 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jll8\" (UniqueName: \"kubernetes.io/projected/0e6edd3c-d469-4a55-a9b8-2a3087248db6-kube-api-access-7jll8\") pod \"collect-profiles-29401110-46chk\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.467591 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.475476 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2pqps"] Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.480047 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.487800 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.493165 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s"] Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.493630 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt5g6\" (UniqueName: \"kubernetes.io/projected/e4449b1d-871a-4134-bc0a-0cbf2508e10b-kube-api-access-rt5g6\") pod \"dns-operator-744455d44c-whvmw\" (UID: \"e4449b1d-871a-4134-bc0a-0cbf2508e10b\") " pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.494460 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.503101 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lccb\" (UniqueName: \"kubernetes.io/projected/6ae04990-84f3-4612-9325-0a40352e1278-kube-api-access-8lccb\") pod \"service-ca-9c57cc56f-h6tc4\" (UID: \"6ae04990-84f3-4612-9325-0a40352e1278\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.506797 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.507102 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.007086024 +0000 UTC m=+142.820702611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.507187 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.510086 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51aacc61-4839-41db-a5ae-a92738581de1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pmsq5\" (UID: \"51aacc61-4839-41db-a5ae-a92738581de1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.530146 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.533125 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.538703 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9r5l\" (UniqueName: \"kubernetes.io/projected/ce31dfaa-5afe-42b6-91e4-635ce4dd3e23-kube-api-access-f9r5l\") pod \"authentication-operator-69f744f599-7kd7r\" (UID: \"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.549584 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7xhqh" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.550209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86ks8\" (UniqueName: \"kubernetes.io/projected/19c82916-91f4-4a0f-9beb-a65cdea01a8a-kube-api-access-86ks8\") pod \"service-ca-operator-777779d784-72gwc\" (UID: \"19c82916-91f4-4a0f-9beb-a65cdea01a8a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.579104 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qhfz\" (UniqueName: \"kubernetes.io/projected/78279129-b25e-4e7c-b563-6043c9ff4f88-kube-api-access-2qhfz\") pod \"marketplace-operator-79b997595-wkn7w\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.579233 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.607803 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.608125 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.108112117 +0000 UTC m=+142.921728704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.650442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvfx\" (UniqueName: \"kubernetes.io/projected/162ad055-0833-461f-92e6-496d1bb33e21-kube-api-access-bhvfx\") pod \"packageserver-d55dfcdfc-szcdp\" (UID: \"162ad055-0833-461f-92e6-496d1bb33e21\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.650999 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ctk8\" (UniqueName: \"kubernetes.io/projected/dc6b507a-226b-42ab-9a86-d4376ba9dfd7-kube-api-access-2ctk8\") pod \"ingress-canary-4pxc7\" (UID: \"dc6b507a-226b-42ab-9a86-d4376ba9dfd7\") " pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.653575 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc9lw\" (UniqueName: \"kubernetes.io/projected/1829a4cf-9696-44fe-a3ec-f3540ac23f81-kube-api-access-tc9lw\") pod \"dns-default-dkss6\" (UID: \"1829a4cf-9696-44fe-a3ec-f3540ac23f81\") " pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.658510 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8jsw\" (UniqueName: \"kubernetes.io/projected/2e1aecfe-07a2-4772-bab2-52b5f5fed397-kube-api-access-q8jsw\") pod \"oauth-openshift-558db77b4-nvzm5\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.708415 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.708639 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.208625427 +0000 UTC m=+143.022242014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.729996 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.765732 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.780413 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.801536 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.820403 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.820618 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.820798 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.320783853 +0000 UTC m=+143.134400440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.840502 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.845261 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-k24kz" event={"ID":"84bba043-a192-4ea4-9225-73cc33b31ea5","Type":"ContainerStarted","Data":"8908be617f4cf6cce955278e7b1ab9286b650dde23e9dc2784c2d61d148a0993"} Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.845974 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" event={"ID":"a0679b82-8f22-4227-bdbf-d92e063ffba0","Type":"ContainerStarted","Data":"a9d0a1c83160c66da897f8c8580b83c6f9e7c476efa2733d76c167107ef7bd06"} Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.846561 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" event={"ID":"36c304aa-55a3-49e5-ab60-394509f4b3ca","Type":"ContainerStarted","Data":"7493ae08084b350b46eca09bb2a26387934817d4b9eaf6262d7dfa04c2d7c33e"} Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.850781 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" event={"ID":"50d299fa-89e9-41ba-8894-af097502b258","Type":"ContainerStarted","Data":"4ad7e9b69f766e513079c57c33fb927d2c2d2197a5a7043237fd14eb6654c3d9"} Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.869648 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.889732 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4pxc7" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.890128 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5g9hp"] Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.925584 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:05 crc kubenswrapper[4696]: E1125 10:39:05.926027 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.426011068 +0000 UTC m=+143.239627655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.926416 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m"] Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.965706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" event={"ID":"5b614b67-512d-474b-ac4e-8cab982cc2f7","Type":"ContainerStarted","Data":"90baf256bd0f008bb805c5a097b8eb30fb575e9fdf46988a6da8c9c1a17e783e"} Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.966868 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.978990 4696 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-clh68 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.979054 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" podUID="0be09268-4279-43e9-8da0-78c4cdd920cd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 25 10:39:05 crc kubenswrapper[4696]: I1125 10:39:05.984703 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.028684 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.031064 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.531051286 +0000 UTC m=+143.344667873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.078123 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg"] Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.078163 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mjb8l"] Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.132386 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.132618 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.632567835 +0000 UTC m=+143.446184432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.132865 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.133275 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.633261297 +0000 UTC m=+143.446877884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.236227 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.236533 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.73651602 +0000 UTC m=+143.550132607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.292413 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48"] Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.342384 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.342843 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.842674883 +0000 UTC m=+143.656291470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.445063 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.445304 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.945276655 +0000 UTC m=+143.758893232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.445344 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.445859 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:06.945851253 +0000 UTC m=+143.759467850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.547113 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.547237 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.047217488 +0000 UTC m=+143.860834075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.547443 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.547894 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.047885909 +0000 UTC m=+143.861502496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.654158 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.654478 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.154462905 +0000 UTC m=+143.968079492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.755985 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.756339 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.256326764 +0000 UTC m=+144.069943351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.862002 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.862396 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.362380804 +0000 UTC m=+144.175997391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:06 crc kubenswrapper[4696]: I1125 10:39:06.964636 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:06 crc kubenswrapper[4696]: E1125 10:39:06.966637 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.466624588 +0000 UTC m=+144.280241175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.000080 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" event={"ID":"23a01a08-d008-451f-86c0-f3c456fafb88","Type":"ContainerStarted","Data":"d31594e1b547b54f3b42f749a8955cf69838ac27b59c78d19f85536054d0daac"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.056079 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4z2br"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.073035 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sbgf7"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.076947 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.077215 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.577200137 +0000 UTC m=+144.390816724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.095516 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mjb8l" event={"ID":"e6d2a426-4cba-4b45-bce3-45633b970e8e","Type":"ContainerStarted","Data":"232c79c4b169a3f0e14717ac3317d3d6c32215f4d4406f6981a7a33a8b28a0d8"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.126836 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" event={"ID":"1695d071-bf51-4bed-a09c-5feefd87adfd","Type":"ContainerStarted","Data":"c833be821889f180121e9f2e152fcce7ee7fc5ca1ba2ae70d5252c2c7b007e21"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.179182 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.182557 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.682541824 +0000 UTC m=+144.496158411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.187823 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7xhqh" event={"ID":"90f9987e-5e7b-40a3-a750-dac2b017942a","Type":"ContainerStarted","Data":"41c3ba645eb4e268b7c7176e5ca0af8744e14b0fb11f25f4373fdae8c1e6dde1"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.195725 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" podStartSLOduration=123.195653638 podStartE2EDuration="2m3.195653638s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.195215405 +0000 UTC m=+144.008831992" watchObservedRunningTime="2025-11-25 10:39:07.195653638 +0000 UTC m=+144.009270225" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.198964 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" event={"ID":"4058138e-66c6-45fe-a684-dc6e7b5b8339","Type":"ContainerStarted","Data":"424db4c1b3e109d755c082d9a313c579709e1073db8ec831c68ef17afc83e01b"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.205936 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dmhj7" podStartSLOduration=124.205914515 podStartE2EDuration="2m4.205914515s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.137763414 +0000 UTC m=+143.951380001" watchObservedRunningTime="2025-11-25 10:39:07.205914515 +0000 UTC m=+144.019531112" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.242600 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" event={"ID":"5b614b67-512d-474b-ac4e-8cab982cc2f7","Type":"ContainerStarted","Data":"a2b33dbe6f00b3adf22aa33de233005af8df0ccdbb0a64b51d58f7985d057b28"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.254009 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" event={"ID":"67e9bbce-9822-49de-934c-0ab922daf87e","Type":"ContainerStarted","Data":"5cd9f3c1e1195987a44a13c417710471a6d8e3adf258703021d88e8d5e903a6f"} Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.280355 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.280870 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.780855445 +0000 UTC m=+144.594472032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.315276 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6b4sd" podStartSLOduration=124.315036879 podStartE2EDuration="2m4.315036879s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.295204637 +0000 UTC m=+144.108821224" watchObservedRunningTime="2025-11-25 10:39:07.315036879 +0000 UTC m=+144.128653466" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.343016 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.383741 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.386691 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.886637796 +0000 UTC m=+144.700254383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.388407 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.390617 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" podStartSLOduration=124.390593648 podStartE2EDuration="2m4.390593648s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.343480926 +0000 UTC m=+144.157097503" watchObservedRunningTime="2025-11-25 10:39:07.390593648 +0000 UTC m=+144.204210235" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.390784 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t5jvs" podStartSLOduration=123.390777614 podStartE2EDuration="2m3.390777614s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.385478391 +0000 UTC m=+144.199094978" watchObservedRunningTime="2025-11-25 10:39:07.390777614 +0000 UTC m=+144.204394221" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.416476 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-whvmw"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.442510 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.471247 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.472391 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.487227 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.487576 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:07.987560207 +0000 UTC m=+144.801176784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.534876 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.534924 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.550546 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" podStartSLOduration=123.550530209 podStartE2EDuration="2m3.550530209s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.548291669 +0000 UTC m=+144.361908256" watchObservedRunningTime="2025-11-25 10:39:07.550530209 +0000 UTC m=+144.364146796" Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.552526 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.554278 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.587791 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp"] Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.588912 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.589271 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.089258203 +0000 UTC m=+144.902874790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: W1125 10:39:07.591084 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f6f8fdc_fec0_46b9_9277_2996da698f24.slice/crio-6cf1282db8de917dfaa2f49a569d1f1bad0d4202cf8caf2673f1257a561bdba2 WatchSource:0}: Error finding container 6cf1282db8de917dfaa2f49a569d1f1bad0d4202cf8caf2673f1257a561bdba2: Status 404 returned error can't find the container with id 6cf1282db8de917dfaa2f49a569d1f1bad0d4202cf8caf2673f1257a561bdba2 Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.594368 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr"] Nov 25 10:39:07 crc kubenswrapper[4696]: W1125 10:39:07.603747 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6243043f_a603_4094_9f48_338241d77432.slice/crio-db970757b8dbc5d35cf91a824d587185bb479acaff171fe6da4c7200e4742898 WatchSource:0}: Error finding container db970757b8dbc5d35cf91a824d587185bb479acaff171fe6da4c7200e4742898: Status 404 returned error can't find the container with id db970757b8dbc5d35cf91a824d587185bb479acaff171fe6da4c7200e4742898 Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.690318 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.691625 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.191610167 +0000 UTC m=+145.005226754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: W1125 10:39:07.700253 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc715c60a_acd3_45e9_81df_98802b303a48.slice/crio-f1081e51d6aa7adbc4c25c11073a20450b02b5097da7141ae09bcf6003a5d1f4 WatchSource:0}: Error finding container f1081e51d6aa7adbc4c25c11073a20450b02b5097da7141ae09bcf6003a5d1f4: Status 404 returned error can't find the container with id f1081e51d6aa7adbc4c25c11073a20450b02b5097da7141ae09bcf6003a5d1f4 Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.792400 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.792756 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.292745645 +0000 UTC m=+145.106362232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.895648 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.896319 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.396303208 +0000 UTC m=+145.209919795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:07 crc kubenswrapper[4696]: I1125 10:39:07.998254 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:07 crc kubenswrapper[4696]: E1125 10:39:07.998558 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.49854665 +0000 UTC m=+145.312163237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.065944 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-k24kz" podStartSLOduration=124.065925226 podStartE2EDuration="2m4.065925226s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:07.950005313 +0000 UTC m=+144.763621900" watchObservedRunningTime="2025-11-25 10:39:08.065925226 +0000 UTC m=+144.879541813" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.099223 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.099778 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.599757459 +0000 UTC m=+145.413374046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.111360 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h6tc4"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.155181 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bv2wn"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.163070 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.200566 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.205450 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.206070 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.706053316 +0000 UTC m=+145.519669903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.306587 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.306744 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.80672898 +0000 UTC m=+145.620345567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.307591 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.307958 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.807949798 +0000 UTC m=+145.621566385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.330207 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5"] Nov 25 10:39:08 crc kubenswrapper[4696]: W1125 10:39:08.342753 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8aa7eb9_5023_45b9_b7b6_10d4feda652c.slice/crio-4ac8cb53e0b39e73d4a1b314aad67797a507d68ddf0c6032cc021caecce957e9 WatchSource:0}: Error finding container 4ac8cb53e0b39e73d4a1b314aad67797a507d68ddf0c6032cc021caecce957e9: Status 404 returned error can't find the container with id 4ac8cb53e0b39e73d4a1b314aad67797a507d68ddf0c6032cc021caecce957e9 Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.344407 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7kd7r"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.398228 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:08 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:08 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:08 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.398278 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.409552 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.411625 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:08.911595463 +0000 UTC m=+145.725212050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.443369 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" event={"ID":"5b614b67-512d-474b-ac4e-8cab982cc2f7","Type":"ContainerStarted","Data":"b90efa287a5269c9e9c5aec64c25e2504b0b3818eccc32fad7f1275cd7b95c04"} Nov 25 10:39:08 crc kubenswrapper[4696]: W1125 10:39:08.454272 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce31dfaa_5afe_42b6_91e4_635ce4dd3e23.slice/crio-eda352e98433b8c4751b332936046c64b15a9187b068b0f01508acf5b50170a2 WatchSource:0}: Error finding container eda352e98433b8c4751b332936046c64b15a9187b068b0f01508acf5b50170a2: Status 404 returned error can't find the container with id eda352e98433b8c4751b332936046c64b15a9187b068b0f01508acf5b50170a2 Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.458562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" event={"ID":"6243043f-a603-4094-9f48-338241d77432","Type":"ContainerStarted","Data":"db970757b8dbc5d35cf91a824d587185bb479acaff171fe6da4c7200e4742898"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.512738 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.529330 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.029312292 +0000 UTC m=+145.842928879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.539621 4696 generic.go:334] "Generic (PLEG): container finished" podID="a0679b82-8f22-4227-bdbf-d92e063ffba0" containerID="2066f278329668572ecc3f5221db87b3b4b192e09ed61d9fd98d733a863e39b8" exitCode=0 Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.539705 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" event={"ID":"a0679b82-8f22-4227-bdbf-d92e063ffba0","Type":"ContainerDied","Data":"2066f278329668572ecc3f5221db87b3b4b192e09ed61d9fd98d733a863e39b8"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.576317 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" event={"ID":"106060a8-1b39-40b0-ac39-b0f057c9552c","Type":"ContainerStarted","Data":"9345f956eab0c51d502b84706a5109363119ae02c97c6d18d72addbb16c55953"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.614836 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.616507 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.11648264 +0000 UTC m=+145.930099227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.674762 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bbdns"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.685750 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" event={"ID":"23a01a08-d008-451f-86c0-f3c456fafb88","Type":"ContainerStarted","Data":"70af6a6e18f7548a441cad1e423caec347dd0ace1b534fcbdb9f80634683e983"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.685799 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" event={"ID":"23a01a08-d008-451f-86c0-f3c456fafb88","Type":"ContainerStarted","Data":"232e62d769434b5431ea7095571726a550de35c630251823fbc736308b331824"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.687530 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkn7w"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.722532 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.722954 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.222940341 +0000 UTC m=+146.036556928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.744339 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mjb8l" event={"ID":"e6d2a426-4cba-4b45-bce3-45633b970e8e","Type":"ContainerStarted","Data":"e096d5c92d632361170b6505a53c13605fbf1628c42fa10d9cbd90f2007a8479"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.745283 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.758952 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.759095 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.781953 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" event={"ID":"c715c60a-acd3-45e9-81df-98802b303a48","Type":"ContainerStarted","Data":"f1081e51d6aa7adbc4c25c11073a20450b02b5097da7141ae09bcf6003a5d1f4"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.789143 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" event={"ID":"6ae04990-84f3-4612-9325-0a40352e1278","Type":"ContainerStarted","Data":"4554f46e462d16a8431577446061ae68fdf70c0fec030943609d3e36d6651bf2"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.792064 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6sm87" podStartSLOduration=124.792049322 podStartE2EDuration="2m4.792049322s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:08.782551739 +0000 UTC m=+145.596168326" watchObservedRunningTime="2025-11-25 10:39:08.792049322 +0000 UTC m=+145.605665909" Nov 25 10:39:08 crc kubenswrapper[4696]: W1125 10:39:08.799347 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod020fd783_0219_4f9d_8504_f430877d9ed4.slice/crio-beca66ff95bf722e77dcbcd8f80f5fba6af83904b150de562ac7adbefc573443 WatchSource:0}: Error finding container beca66ff95bf722e77dcbcd8f80f5fba6af83904b150de562ac7adbefc573443: Status 404 returned error can't find the container with id beca66ff95bf722e77dcbcd8f80f5fba6af83904b150de562ac7adbefc573443 Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.812316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" event={"ID":"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9","Type":"ContainerStarted","Data":"7bbf9ebf3b949b961157d9b9021d8f001889ef98652d930b506ceae0c6628dc1"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.823352 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.824799 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.32478015 +0000 UTC m=+146.138396737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.833910 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" event={"ID":"4058138e-66c6-45fe-a684-dc6e7b5b8339","Type":"ContainerStarted","Data":"5805c1a1dc262030822d9b0dca8510ce77c02578f570aab421e4c303be8c1bbe"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.834756 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.845719 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4pxc7"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.846179 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5g9hp" podStartSLOduration=124.84615881 podStartE2EDuration="2m4.84615881s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:08.838902616 +0000 UTC m=+145.652519203" watchObservedRunningTime="2025-11-25 10:39:08.84615881 +0000 UTC m=+145.659775397" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.849425 4696 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-qwr48 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.849639 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" podUID="4058138e-66c6-45fe-a684-dc6e7b5b8339" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.878205 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" event={"ID":"0af4515c-aa04-454f-ac8c-412ba49f295a","Type":"ContainerStarted","Data":"fb163eca1fb506a649574eab064cffe7bd44fd001f6636b3018eb18c0eb124d3"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.891926 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rj85t"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.904860 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" event={"ID":"8a4561e0-205d-4866-8040-c3cec49d3127","Type":"ContainerStarted","Data":"c1111b064ca87c6cb070d2bbc738b5ebfee8e4e70c809e8db09e252bb844de0d"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.907421 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mjb8l" podStartSLOduration=125.907391338 podStartE2EDuration="2m5.907391338s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:08.89644963 +0000 UTC m=+145.710066217" watchObservedRunningTime="2025-11-25 10:39:08.907391338 +0000 UTC m=+145.721007925" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.909336 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-72gwc"] Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.925221 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.926656 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" podStartSLOduration=124.926635861 podStartE2EDuration="2m4.926635861s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:08.924410562 +0000 UTC m=+145.738027149" watchObservedRunningTime="2025-11-25 10:39:08.926635861 +0000 UTC m=+145.740252448" Nov 25 10:39:08 crc kubenswrapper[4696]: E1125 10:39:08.928145 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.428126847 +0000 UTC m=+146.241743434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.951371 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" event={"ID":"0330c6a6-2aaf-4621-9983-f947a3bacde9","Type":"ContainerStarted","Data":"ef5626d51a75303bef193f68d905072a24f4dec8369c9a9edc741595630b1b20"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.951432 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" event={"ID":"0330c6a6-2aaf-4621-9983-f947a3bacde9","Type":"ContainerStarted","Data":"e7312e9c4f99401c6dcb35492024538f5031fb9ba3d7db8d7a95074d04eabba0"} Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.952497 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.954916 4696 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbgf7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 10:39:08 crc kubenswrapper[4696]: I1125 10:39:08.954942 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" podUID="0330c6a6-2aaf-4621-9983-f947a3bacde9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.032625 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.034408 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.534195137 +0000 UTC m=+146.347811724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.036186 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.037068 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-k24kz" event={"ID":"84bba043-a192-4ea4-9225-73cc33b31ea5","Type":"ContainerStarted","Data":"6b9087db56fbe1ff0c492c3744a92c557667b07776680108be92f0db2c6cbb8d"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.043495 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" event={"ID":"bbc2f0e2-9d49-4d3c-9c44-15b762505b83","Type":"ContainerStarted","Data":"6bb08ef07223c24719ed3b5fdaab61cdb6e0426aae5dc436e6fe306801896cea"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.050085 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dkss6"] Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.051010 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.550992255 +0000 UTC m=+146.364608832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.062326 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk"] Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.069618 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7xhqh" event={"ID":"90f9987e-5e7b-40a3-a750-dac2b017942a","Type":"ContainerStarted","Data":"f3da848b7db449b972479a471acf741534dc6a42daa6454059a493aaab56d464"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.088051 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nvzm5"] Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.107682 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" podStartSLOduration=126.1076448 podStartE2EDuration="2m6.1076448s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:09.055881125 +0000 UTC m=+145.869497722" watchObservedRunningTime="2025-11-25 10:39:09.1076448 +0000 UTC m=+145.921261387" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.108760 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" event={"ID":"2f6f8fdc-fec0-46b9-9277-2996da698f24","Type":"ContainerStarted","Data":"6cf1282db8de917dfaa2f49a569d1f1bad0d4202cf8caf2673f1257a561bdba2"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.116778 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7xhqh" podStartSLOduration=7.116737231 podStartE2EDuration="7.116737231s" podCreationTimestamp="2025-11-25 10:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:09.108464866 +0000 UTC m=+145.922081453" watchObservedRunningTime="2025-11-25 10:39:09.116737231 +0000 UTC m=+145.930353818" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.141834 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.143029 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.643007951 +0000 UTC m=+146.456624548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.143552 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" event={"ID":"84d56ecc-4b70-4776-b906-4228569ddac6","Type":"ContainerStarted","Data":"f442e4335a5f69eab9219bcc75ffa7c2230d767e23bd1bf7e5b93ceefc1a4665"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.143628 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.144054 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:09 crc kubenswrapper[4696]: W1125 10:39:09.151780 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19c82916_91f4_4a0f_9beb_a65cdea01a8a.slice/crio-1154dfb32c46b69e23f9183ae599ee871750e126bbc18fd01a658dcf40f8bda0 WatchSource:0}: Error finding container 1154dfb32c46b69e23f9183ae599ee871750e126bbc18fd01a658dcf40f8bda0: Status 404 returned error can't find the container with id 1154dfb32c46b69e23f9183ae599ee871750e126bbc18fd01a658dcf40f8bda0 Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.157366 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" podStartSLOduration=125.157348704 podStartE2EDuration="2m5.157348704s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:09.157100616 +0000 UTC m=+145.970717203" watchObservedRunningTime="2025-11-25 10:39:09.157348704 +0000 UTC m=+145.970965291" Nov 25 10:39:09 crc kubenswrapper[4696]: W1125 10:39:09.159015 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1829a4cf_9696_44fe_a3ec_f3540ac23f81.slice/crio-ea9871d52d8ac7ae3e306921605647d20f5af430639dc4d2df4f501b470fcba2 WatchSource:0}: Error finding container ea9871d52d8ac7ae3e306921605647d20f5af430639dc4d2df4f501b470fcba2: Status 404 returned error can't find the container with id ea9871d52d8ac7ae3e306921605647d20f5af430639dc4d2df4f501b470fcba2 Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.170974 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" event={"ID":"67e9bbce-9822-49de-934c-0ab922daf87e","Type":"ContainerStarted","Data":"9948fdcaa6a6521e30285eb5d09f8504e6f350e673c0c3cbc9468e7b3087d03d"} Nov 25 10:39:09 crc kubenswrapper[4696]: W1125 10:39:09.179350 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e6edd3c_d469_4a55_a9b8_2a3087248db6.slice/crio-bc7cac1de148f4c3688cf6a1bf7bda623c61e2347f3554d52704f1ac383b2ad3 WatchSource:0}: Error finding container bc7cac1de148f4c3688cf6a1bf7bda623c61e2347f3554d52704f1ac383b2ad3: Status 404 returned error can't find the container with id bc7cac1de148f4c3688cf6a1bf7bda623c61e2347f3554d52704f1ac383b2ad3 Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.185952 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.190508 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" event={"ID":"1695d071-bf51-4bed-a09c-5feefd87adfd","Type":"ContainerStarted","Data":"25bf5c8ad778624ba9fa0e467935bd98785124f9f023ebc8b458b304507bf1e0"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.190554 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" event={"ID":"1695d071-bf51-4bed-a09c-5feefd87adfd","Type":"ContainerStarted","Data":"91c84c699e008fcc332b10c55e2c16ace9b4d19417220b12a3912833847197b3"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.191458 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.233380 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-s4kcg" podStartSLOduration=125.233361316 podStartE2EDuration="2m5.233361316s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:09.191191107 +0000 UTC m=+146.004807694" watchObservedRunningTime="2025-11-25 10:39:09.233361316 +0000 UTC m=+146.046977903" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.242114 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" event={"ID":"e4449b1d-871a-4134-bc0a-0cbf2508e10b","Type":"ContainerStarted","Data":"d557f8fb7f0a161646690d31c021f89004271aff69371f704dce297475108977"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.243630 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.245579 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.745564372 +0000 UTC m=+146.559180959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.274429 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" event={"ID":"36c304aa-55a3-49e5-ab60-394509f4b3ca","Type":"ContainerStarted","Data":"e4dcfbcd6fe13b5205fc4ccaaeac054314a1450c2223c74f5996bb363d43bc91"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.274457 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" event={"ID":"36c304aa-55a3-49e5-ab60-394509f4b3ca","Type":"ContainerStarted","Data":"31afd51d02238bbb519acb58543a33bab836c6dc2f9dee47fbb74c920d24fc60"} Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.290447 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mssfq" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.307614 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" podStartSLOduration=125.307592245 podStartE2EDuration="2m5.307592245s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:09.306855122 +0000 UTC m=+146.120471709" watchObservedRunningTime="2025-11-25 10:39:09.307592245 +0000 UTC m=+146.121208832" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.312293 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp"] Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.345331 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.346347 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.846332659 +0000 UTC m=+146.659949246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.394834 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:09 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:09 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:09 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.395241 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.449961 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.455109 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:09.955094542 +0000 UTC m=+146.768711129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.470678 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkt6s" podStartSLOduration=125.470633231 podStartE2EDuration="2m5.470633231s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:09.395709292 +0000 UTC m=+146.209325889" watchObservedRunningTime="2025-11-25 10:39:09.470633231 +0000 UTC m=+146.284249818" Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.550642 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.550809 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.050784482 +0000 UTC m=+146.864401069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.551572 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.551853 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.051839355 +0000 UTC m=+146.865455942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.652941 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.653314 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.153300172 +0000 UTC m=+146.966916759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.756158 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.756590 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.256579527 +0000 UTC m=+147.070196114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.856980 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.857353 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.357336603 +0000 UTC m=+147.170953190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:09 crc kubenswrapper[4696]: I1125 10:39:09.958994 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:09 crc kubenswrapper[4696]: E1125 10:39:09.959865 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.459853683 +0000 UTC m=+147.273470270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.062973 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.063309 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.563282451 +0000 UTC m=+147.376899038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.166403 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.167383 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.66736938 +0000 UTC m=+147.480985967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.268620 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.268899 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.76888445 +0000 UTC m=+147.582501027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.289041 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" event={"ID":"a0679b82-8f22-4227-bdbf-d92e063ffba0","Type":"ContainerStarted","Data":"394297628ec5473dd442ef4118c520cff8392cdcaebbf5aff28bdb4b587a3d73"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.289779 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.299301 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" event={"ID":"c715c60a-acd3-45e9-81df-98802b303a48","Type":"ContainerStarted","Data":"b68c4d551aa58b443a75c8d4961652c7fe20b17ab58a3d6b00c43333aa69d889"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.300484 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.337554 4696 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-lpgdr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.337605 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" podUID="c715c60a-acd3-45e9-81df-98802b303a48" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.349726 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" event={"ID":"8a4561e0-205d-4866-8040-c3cec49d3127","Type":"ContainerStarted","Data":"444d434ed55e794d8cc5e5a915501b182ee9b4e379ace678687e0cb1f61d2c87"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.356506 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" podStartSLOduration=127.35647855 podStartE2EDuration="2m7.35647855s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.329272001 +0000 UTC m=+147.142888588" watchObservedRunningTime="2025-11-25 10:39:10.35647855 +0000 UTC m=+147.170095137" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.369988 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.370627 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.870607265 +0000 UTC m=+147.684223852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.393802 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" podStartSLOduration=126.39378978 podStartE2EDuration="2m6.39378978s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.39377677 +0000 UTC m=+147.207393357" watchObservedRunningTime="2025-11-25 10:39:10.39378978 +0000 UTC m=+147.207406367" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.396050 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" event={"ID":"106060a8-1b39-40b0-ac39-b0f057c9552c","Type":"ContainerStarted","Data":"d2e76df953b414665aea6bd8a2af52978099d53f97af26dcf1bc1c25e559a9c7"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.396098 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" event={"ID":"106060a8-1b39-40b0-ac39-b0f057c9552c","Type":"ContainerStarted","Data":"4216c29684da5d67fa0e9c4b86fddc2b1c32c6702f4bca8766bffaade14c4f7c"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.400171 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:10 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:10 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:10 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.400221 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.401438 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" event={"ID":"19c82916-91f4-4a0f-9beb-a65cdea01a8a","Type":"ContainerStarted","Data":"6757e18de445c69a6147f1e440494a194ed2ce89a67d24092970f54b29a5b288"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.401473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" event={"ID":"19c82916-91f4-4a0f-9beb-a65cdea01a8a","Type":"ContainerStarted","Data":"1154dfb32c46b69e23f9183ae599ee871750e126bbc18fd01a658dcf40f8bda0"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.408271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rj85t" event={"ID":"376ea9c4-9e68-4475-b205-95e2185e51ba","Type":"ContainerStarted","Data":"1d793054554ae20fae52561dbb6c5ee7c3a0fea6bb62172540aea264b9b1bf2b"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.408318 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rj85t" event={"ID":"376ea9c4-9e68-4475-b205-95e2185e51ba","Type":"ContainerStarted","Data":"1e2cd1359cb85fa71f7451b4377204d89df1d5c5a90787309f387ad09c232586"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.411809 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" event={"ID":"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23","Type":"ContainerStarted","Data":"eda352e98433b8c4751b332936046c64b15a9187b068b0f01508acf5b50170a2"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.416681 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dkss6" event={"ID":"1829a4cf-9696-44fe-a3ec-f3540ac23f81","Type":"ContainerStarted","Data":"ea9871d52d8ac7ae3e306921605647d20f5af430639dc4d2df4f501b470fcba2"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.426923 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-4z2br" podStartSLOduration=126.4268912 podStartE2EDuration="2m6.4268912s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.426290682 +0000 UTC m=+147.239907269" watchObservedRunningTime="2025-11-25 10:39:10.4268912 +0000 UTC m=+147.240507777" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.433954 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" event={"ID":"2e1aecfe-07a2-4772-bab2-52b5f5fed397","Type":"ContainerStarted","Data":"8fa15718bc52862248ccf9428cd3fa22641d1f3c619fbf39b18188f9b744da76"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.446870 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" event={"ID":"6ae04990-84f3-4612-9325-0a40352e1278","Type":"ContainerStarted","Data":"2672af28e436b1768ca5fd944f2270916cd4ccfc52be0778bbce54e45be31e22"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.467110 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-72gwc" podStartSLOduration=126.46709189 podStartE2EDuration="2m6.46709189s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.464475939 +0000 UTC m=+147.278092526" watchObservedRunningTime="2025-11-25 10:39:10.46709189 +0000 UTC m=+147.280708477" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.472222 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.474079 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:10.974058404 +0000 UTC m=+147.787674981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.482495 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" event={"ID":"f8aa7eb9-5023-45b9-b7b6-10d4feda652c","Type":"ContainerStarted","Data":"94467c56eeed6400fd0e903cde7f0d9dbf54d51cedd7f8388e5868e7e37ce092"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.482554 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" event={"ID":"f8aa7eb9-5023-45b9-b7b6-10d4feda652c","Type":"ContainerStarted","Data":"4ac8cb53e0b39e73d4a1b314aad67797a507d68ddf0c6032cc021caecce957e9"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.490857 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" event={"ID":"0e6edd3c-d469-4a55-a9b8-2a3087248db6","Type":"ContainerStarted","Data":"bc7cac1de148f4c3688cf6a1bf7bda623c61e2347f3554d52704f1ac383b2ad3"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.497943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" event={"ID":"78279129-b25e-4e7c-b563-6043c9ff4f88","Type":"ContainerStarted","Data":"e10f3008b1d2381f305ebce42392850e08eeb28679e31bc8c71d01a47cac3d67"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.518878 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk5fp" podStartSLOduration=127.518862405 podStartE2EDuration="2m7.518862405s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.51674279 +0000 UTC m=+147.330359387" watchObservedRunningTime="2025-11-25 10:39:10.518862405 +0000 UTC m=+147.332478992" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.519622 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n62qw" event={"ID":"2f6f8fdc-fec0-46b9-9277-2996da698f24","Type":"ContainerStarted","Data":"9018ab4c0fe83e2cf389fefbe6abf42961c4ff2d63bf98b96317aa6da8b64f08"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.544141 4696 generic.go:334] "Generic (PLEG): container finished" podID="29a2d684-8d98-42c3-9d91-12d0a87bf032" containerID="cd7a1369b87dc92ca2c7efd0c62c707a828db0d074f926ecce8ca33112a5b62d" exitCode=0 Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.544315 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" event={"ID":"29a2d684-8d98-42c3-9d91-12d0a87bf032","Type":"ContainerDied","Data":"cd7a1369b87dc92ca2c7efd0c62c707a828db0d074f926ecce8ca33112a5b62d"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.544361 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" event={"ID":"29a2d684-8d98-42c3-9d91-12d0a87bf032","Type":"ContainerStarted","Data":"da3a3ec95b394e7e218a94a47df2d447522af38e5fe94963ddd0f1b79867a63e"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.559970 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" event={"ID":"0af4515c-aa04-454f-ac8c-412ba49f295a","Type":"ContainerStarted","Data":"110b14efff9232f1694c043903f698e248ba17feba6cfd39a1663b71bab55775"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.570584 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-rj85t" podStartSLOduration=127.57056712 podStartE2EDuration="2m7.57056712s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.569228488 +0000 UTC m=+147.382845075" watchObservedRunningTime="2025-11-25 10:39:10.57056712 +0000 UTC m=+147.384183707" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.585881 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.587613 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.087601135 +0000 UTC m=+147.901217722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.604979 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" event={"ID":"6243043f-a603-4094-9f48-338241d77432","Type":"ContainerStarted","Data":"8f32da467ab53f4b2ca85357153fff075e141c736813e9cd1174d6f8fab9a13d"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.605029 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" event={"ID":"6243043f-a603-4094-9f48-338241d77432","Type":"ContainerStarted","Data":"7d16f0958f5bec14bf3f056c4529b4dc61d5ded7d2f0555b1aea1083602aae86"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.629931 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" event={"ID":"162ad055-0833-461f-92e6-496d1bb33e21","Type":"ContainerStarted","Data":"a2d7c70ed675f6a1903d0c71493161b93b5e11161cb4adc380e4f4cc7a65ead5"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.637858 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" event={"ID":"bbc2f0e2-9d49-4d3c-9c44-15b762505b83","Type":"ContainerStarted","Data":"3c5e994bd20a35b90a56cbd6528d355a301ad2f3ec7e13b2a54244ed012e5661"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.653912 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" event={"ID":"020fd783-0219-4f9d-8504-f430877d9ed4","Type":"ContainerStarted","Data":"beca66ff95bf722e77dcbcd8f80f5fba6af83904b150de562ac7adbefc573443"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.664018 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" event={"ID":"51aacc61-4839-41db-a5ae-a92738581de1","Type":"ContainerStarted","Data":"4ee1d6b3e4096e974089092cc71458e35d29f9f5703b414803593fc6d2c5bb77"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.664078 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" event={"ID":"51aacc61-4839-41db-a5ae-a92738581de1","Type":"ContainerStarted","Data":"0630c750e4156a5bf5f23373af15b9ed315ed78d2ecc24978630fbb74c0c812b"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.670473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4pxc7" event={"ID":"dc6b507a-226b-42ab-9a86-d4376ba9dfd7","Type":"ContainerStarted","Data":"389b4437294534d59fa124f2a7ae05f6a12fcabfca40c8d1c351ba845249febe"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.675128 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" event={"ID":"e6b216df-ef9d-4e5a-a92c-f42c5dfb0ca9","Type":"ContainerStarted","Data":"5b705126db4fe3e1441be4637c1ae37fe0ece941d77b7e9f074292cf7c633d7a"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.682964 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" event={"ID":"e4449b1d-871a-4134-bc0a-0cbf2508e10b","Type":"ContainerStarted","Data":"ab782ee51753e9b97a0233f95321112221806c9b7b44fefffa546e372946f361"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.687734 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" event={"ID":"84d56ecc-4b70-4776-b906-4228569ddac6","Type":"ContainerStarted","Data":"1325ef0e824bbdfefc814d0fa646213d8330db030f766b05efda90985f2c43cd"} Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.689418 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.691188 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.191167228 +0000 UTC m=+148.004783815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.691256 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.693740 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.193728847 +0000 UTC m=+148.007345434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.695633 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.695687 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.724755 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qwr48" Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.793204 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.793392 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.293365978 +0000 UTC m=+148.106982565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.793617 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.795247 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.295235856 +0000 UTC m=+148.108852443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.895127 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.895290 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.39526344 +0000 UTC m=+148.208880027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.895416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:10 crc kubenswrapper[4696]: E1125 10:39:10.895766 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.395744464 +0000 UTC m=+148.209361082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:10 crc kubenswrapper[4696]: I1125 10:39:10.996187 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.000157 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.500121371 +0000 UTC m=+148.313737998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.036612 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-h6tc4" podStartSLOduration=127.036591986 podStartE2EDuration="2m7.036591986s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:10.867944978 +0000 UTC m=+147.681561555" watchObservedRunningTime="2025-11-25 10:39:11.036591986 +0000 UTC m=+147.850208573" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.100365 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.100959 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.60094741 +0000 UTC m=+148.414563997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.158235 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4sg27" podStartSLOduration=127.158215646 podStartE2EDuration="2m7.158215646s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.050627329 +0000 UTC m=+147.864243926" watchObservedRunningTime="2025-11-25 10:39:11.158215646 +0000 UTC m=+147.971832233" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.202063 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.202447 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.702424988 +0000 UTC m=+148.516041575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.281395 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wkz7" podStartSLOduration=127.281376192 podStartE2EDuration="2m7.281376192s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.157968578 +0000 UTC m=+147.971585165" watchObservedRunningTime="2025-11-25 10:39:11.281376192 +0000 UTC m=+148.094992779" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.282496 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6dbht" podStartSLOduration=127.282492697 podStartE2EDuration="2m7.282492697s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.281000831 +0000 UTC m=+148.094617418" watchObservedRunningTime="2025-11-25 10:39:11.282492697 +0000 UTC m=+148.096109284" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.303730 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.304167 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.804155865 +0000 UTC m=+148.617772452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.391573 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:11 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:11 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:11 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.391626 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.405092 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.405189 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.905173839 +0000 UTC m=+148.718790426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.405512 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.405769 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:11.905761727 +0000 UTC m=+148.719378314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.507044 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.507434 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.007419561 +0000 UTC m=+148.821036148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.579226 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pmsq5" podStartSLOduration=127.579196494 podStartE2EDuration="2m7.579196494s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.578009437 +0000 UTC m=+148.391626024" watchObservedRunningTime="2025-11-25 10:39:11.579196494 +0000 UTC m=+148.392813081" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.579998 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-g5n86" podStartSLOduration=127.579974827 podStartE2EDuration="2m7.579974827s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.50639413 +0000 UTC m=+148.320010707" watchObservedRunningTime="2025-11-25 10:39:11.579974827 +0000 UTC m=+148.393591414" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.616475 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.633099 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.133077914 +0000 UTC m=+148.946694501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.690345 4696 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbgf7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.690392 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" podUID="0330c6a6-2aaf-4621-9983-f947a3bacde9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.712919 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" event={"ID":"29a2d684-8d98-42c3-9d91-12d0a87bf032","Type":"ContainerStarted","Data":"6f77e0c413d0c11063e2525c3086c90c085c869fbfb6b9b6a68e18b8a87559b9"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.723785 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4pxc7" event={"ID":"dc6b507a-226b-42ab-9a86-d4376ba9dfd7","Type":"ContainerStarted","Data":"05877d4b752d6d2656c686807dd373aa91c52adc4fcabb3edf1b8f45200e239a"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.733244 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.733710 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.233689886 +0000 UTC m=+149.047306473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.748980 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" event={"ID":"ce31dfaa-5afe-42b6-91e4-635ce4dd3e23","Type":"ContainerStarted","Data":"50631bf624e47c4e6342a8e49f0b4f3cfc4cb801f22ee005ed1e713cb3e2316e"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.757129 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4pxc7" podStartSLOduration=9.757112168 podStartE2EDuration="9.757112168s" podCreationTimestamp="2025-11-25 10:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.756949403 +0000 UTC m=+148.570565990" watchObservedRunningTime="2025-11-25 10:39:11.757112168 +0000 UTC m=+148.570728745" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.774075 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" event={"ID":"2e1aecfe-07a2-4772-bab2-52b5f5fed397","Type":"ContainerStarted","Data":"781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.774892 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.778727 4696 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nvzm5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.778797 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.785249 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-7kd7r" podStartSLOduration=127.785228405 podStartE2EDuration="2m7.785228405s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.781709296 +0000 UTC m=+148.595325893" watchObservedRunningTime="2025-11-25 10:39:11.785228405 +0000 UTC m=+148.598844992" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.802177 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dkss6" event={"ID":"1829a4cf-9696-44fe-a3ec-f3540ac23f81","Type":"ContainerStarted","Data":"6dc821f023d6cf54c3d44145eb9180c096fe22cc03fe55ef1cbb92af99b6ff86"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.802222 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dkss6" event={"ID":"1829a4cf-9696-44fe-a3ec-f3540ac23f81","Type":"ContainerStarted","Data":"fd1672df6e94a5be8f2f205ab5e593dc4482b36ead655849878a6ed94d3f3cf5"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.802767 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.808914 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" event={"ID":"f8aa7eb9-5023-45b9-b7b6-10d4feda652c","Type":"ContainerStarted","Data":"f89312cd52d80aaec1062795573fcbdc762d0112f666b3dac3671c6ddf1292e2"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.810615 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" event={"ID":"0e6edd3c-d469-4a55-a9b8-2a3087248db6","Type":"ContainerStarted","Data":"9881f9e2a38942d186ed0835edac34e728e29c17093199894b1b465a1db89cc8"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.811851 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" event={"ID":"78279129-b25e-4e7c-b563-6043c9ff4f88","Type":"ContainerStarted","Data":"26b98ed90897186f3e1061344ddb4bbbf62c8304a33fd262ac83b0d6675c6855"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.814787 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.824830 4696 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wkn7w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.824885 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.837573 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.843085 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.343069858 +0000 UTC m=+149.156686445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.846210 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" event={"ID":"0af4515c-aa04-454f-ac8c-412ba49f295a","Type":"ContainerStarted","Data":"55dec0ca2499e36bf06cd1dd8381038ce5c79ca368e62b9444d582668da2f827"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.866739 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" event={"ID":"162ad055-0833-461f-92e6-496d1bb33e21","Type":"ContainerStarted","Data":"5671a161212a2bbe83c9c4bb0876c2aa61a55743ff890bc457456472c3bab1a8"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.867723 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.872197 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" podStartSLOduration=127.872180806 podStartE2EDuration="2m7.872180806s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.870129363 +0000 UTC m=+148.683745960" watchObservedRunningTime="2025-11-25 10:39:11.872180806 +0000 UTC m=+148.685797413" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.874159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" event={"ID":"020fd783-0219-4f9d-8504-f430877d9ed4","Type":"ContainerStarted","Data":"fb7df1b49793abfb77742d5e119746f21fadf95a340cb8b93117eb6106630e34"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.877845 4696 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-szcdp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.877909 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" podUID="162ad055-0833-461f-92e6-496d1bb33e21" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.891713 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" event={"ID":"e4449b1d-871a-4134-bc0a-0cbf2508e10b","Type":"ContainerStarted","Data":"c0439876773210189eadbf8fc80a43ec3d908d65dd9ac8032827108b883d7f4e"} Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.895441 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.895490 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.896545 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dkss6" podStartSLOduration=9.896535296 podStartE2EDuration="9.896535296s" podCreationTimestamp="2025-11-25 10:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.895044031 +0000 UTC m=+148.708660628" watchObservedRunningTime="2025-11-25 10:39:11.896535296 +0000 UTC m=+148.710151883" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.929632 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b82sf" podStartSLOduration=127.929604245 podStartE2EDuration="2m7.929604245s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.926950614 +0000 UTC m=+148.740567201" watchObservedRunningTime="2025-11-25 10:39:11.929604245 +0000 UTC m=+148.743220832" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.939100 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.940586 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.940736 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.941185 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.941272 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.949789 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:39:11 crc kubenswrapper[4696]: E1125 10:39:11.954354 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.454312148 +0000 UTC m=+149.267928735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.956297 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.962284 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.972057 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.977323 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lpgdr" Nov 25 10:39:11 crc kubenswrapper[4696]: I1125 10:39:11.981743 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:11.998033 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.013488 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.057599 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podStartSLOduration=128.057583182 podStartE2EDuration="2m8.057583182s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:11.998595123 +0000 UTC m=+148.812211710" watchObservedRunningTime="2025-11-25 10:39:12.057583182 +0000 UTC m=+148.871199769" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.058317 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.058649 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.558636863 +0000 UTC m=+149.372253450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.100209 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" podStartSLOduration=128.100194624 podStartE2EDuration="2m8.100194624s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:12.057680335 +0000 UTC m=+148.871296922" watchObservedRunningTime="2025-11-25 10:39:12.100194624 +0000 UTC m=+148.913811211" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.142258 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vt4qt" podStartSLOduration=128.1422353 podStartE2EDuration="2m8.1422353s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:12.140881409 +0000 UTC m=+148.954497996" watchObservedRunningTime="2025-11-25 10:39:12.1422353 +0000 UTC m=+148.955851887" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.142510 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" podStartSLOduration=129.142505019 podStartE2EDuration="2m9.142505019s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:12.101090452 +0000 UTC m=+148.914707029" watchObservedRunningTime="2025-11-25 10:39:12.142505019 +0000 UTC m=+148.956121606" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.162414 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.162838 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.662807945 +0000 UTC m=+149.476424532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.221374 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-whvmw" podStartSLOduration=128.22135835 podStartE2EDuration="2m8.22135835s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:12.219519883 +0000 UTC m=+149.033136470" watchObservedRunningTime="2025-11-25 10:39:12.22135835 +0000 UTC m=+149.034974937" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.264492 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.264835 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.764821 +0000 UTC m=+149.578437587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.366246 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.366528 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.866514235 +0000 UTC m=+149.680130822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.399408 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:12 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:12 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:12 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.399457 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.468177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.468733 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:12.968716476 +0000 UTC m=+149.782333063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.569726 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.570008 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.069993188 +0000 UTC m=+149.883609775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.670654 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.671025 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.171009262 +0000 UTC m=+149.984625849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.706509 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sbgf7" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.771695 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.772058 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.272031886 +0000 UTC m=+150.085648463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.772124 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.772514 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.27249422 +0000 UTC m=+150.086110807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.872937 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.873104 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.373079381 +0000 UTC m=+150.186695968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.873555 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:12 crc kubenswrapper[4696]: E1125 10:39:12.873878 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.373864545 +0000 UTC m=+150.187481132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.908101 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" event={"ID":"29a2d684-8d98-42c3-9d91-12d0a87bf032","Type":"ContainerStarted","Data":"f73d45e640a909215b6dc1a44133c3a305e7d64d08f45b973213ef279b3924ae"} Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.908326 4696 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wkn7w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.908375 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.908672 4696 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nvzm5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.908691 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.975357 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:12 crc kubenswrapper[4696]: I1125 10:39:12.978279 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" podStartSLOduration=128.978264624 podStartE2EDuration="2m8.978264624s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:12.977991556 +0000 UTC m=+149.791608163" watchObservedRunningTime="2025-11-25 10:39:12.978264624 +0000 UTC m=+149.791881211" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:12.997817 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.497785576 +0000 UTC m=+150.311402163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.038061 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nd9j4"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.038995 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.044584 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.079505 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.079910 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.579898488 +0000 UTC m=+150.393515075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.117320 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nd9j4"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.184167 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.184306 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.684285155 +0000 UTC m=+150.497901742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.184747 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5hnq\" (UniqueName: \"kubernetes.io/projected/d6b24f5f-89c0-4078-be5d-8739b554e76e-kube-api-access-z5hnq\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.184779 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-utilities\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.184857 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.184878 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-catalog-content\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.185156 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.685147932 +0000 UTC m=+150.498764519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.270807 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tpmx6"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.272012 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.286207 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.786176616 +0000 UTC m=+150.599793203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.286417 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.286781 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrcx5\" (UniqueName: \"kubernetes.io/projected/46840e91-cc38-4a5c-beca-96f3735590c4-kube-api-access-qrcx5\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.286918 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.287154 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-catalog-content\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.287268 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-catalog-content\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.287331 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.787320141 +0000 UTC m=+150.600936748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.287439 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5hnq\" (UniqueName: \"kubernetes.io/projected/d6b24f5f-89c0-4078-be5d-8739b554e76e-kube-api-access-z5hnq\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.287553 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-utilities\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.287632 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-utilities\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.287509 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-catalog-content\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.288052 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-utilities\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.298747 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.301844 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tpmx6"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.369404 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5hnq\" (UniqueName: \"kubernetes.io/projected/d6b24f5f-89c0-4078-be5d-8739b554e76e-kube-api-access-z5hnq\") pod \"certified-operators-nd9j4\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.387168 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.388282 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.388881 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.888860742 +0000 UTC m=+150.702477329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.388967 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-utilities\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.389031 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrcx5\" (UniqueName: \"kubernetes.io/projected/46840e91-cc38-4a5c-beca-96f3735590c4-kube-api-access-qrcx5\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.389104 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.389138 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-catalog-content\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.389582 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-catalog-content\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.389854 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-utilities\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.390419 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:13.89041087 +0000 UTC m=+150.704027457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.407931 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:13 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:13 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:13 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.408012 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.444434 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrcx5\" (UniqueName: \"kubernetes.io/projected/46840e91-cc38-4a5c-beca-96f3735590c4-kube-api-access-qrcx5\") pod \"community-operators-tpmx6\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.480815 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-88drx"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.498742 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.510682 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.511012 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-catalog-content\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.511051 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-utilities\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.511112 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v942r\" (UniqueName: \"kubernetes.io/projected/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-kube-api-access-v942r\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.511248 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.011230794 +0000 UTC m=+150.824847381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.593497 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.605959 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-88drx"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.616978 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-catalog-content\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.617390 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-utilities\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.617595 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.617722 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v942r\" (UniqueName: \"kubernetes.io/projected/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-kube-api-access-v942r\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.619335 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-catalog-content\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.619658 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.119641777 +0000 UTC m=+150.933258364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.620346 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-utilities\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.661943 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v942r\" (UniqueName: \"kubernetes.io/projected/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-kube-api-access-v942r\") pod \"certified-operators-88drx\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.709855 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jvmp4"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.719490 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.719767 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.219753362 +0000 UTC m=+151.033369939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.720773 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.732352 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvmp4"] Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.820909 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-catalog-content\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.821245 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-utilities\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.821336 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.821425 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s4hj\" (UniqueName: \"kubernetes.io/projected/7df8bce2-d9fd-4875-8af0-86942949d93d-kube-api-access-5s4hj\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.821857 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.321840889 +0000 UTC m=+151.135457476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.848397 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.908585 4696 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-szcdp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.908648 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" podUID="162ad055-0833-461f-92e6-496d1bb33e21" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.923775 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.923921 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-catalog-content\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.923969 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-utilities\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.924031 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s4hj\" (UniqueName: \"kubernetes.io/projected/7df8bce2-d9fd-4875-8af0-86942949d93d-kube-api-access-5s4hj\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: E1125 10:39:13.924383 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.4243687 +0000 UTC m=+151.237985287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.924892 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-catalog-content\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.925112 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-utilities\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.934502 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6c44a535e7a44d1ae68d39d0f2edb8e60fdd41e96f9ca7869393367ab89dc284"} Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.949758 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s4hj\" (UniqueName: \"kubernetes.io/projected/7df8bce2-d9fd-4875-8af0-86942949d93d-kube-api-access-5s4hj\") pod \"community-operators-jvmp4\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.969565 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"293131ccb50d3b45fc2bfa4ac9061eb934c5b42a2171383ab08316e79a80989e"} Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.978764 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0731ab7f43858d0a88b277d643bc86323a19a4d8244ba542a381db7e1a1da07e"} Nov 25 10:39:13 crc kubenswrapper[4696]: I1125 10:39:13.978864 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"344580fb6a23bf01619af80c9cd17682b318994b20624c8ce1692aa20574cc06"} Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.023763 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.024614 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.026448 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.026812 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.526798228 +0000 UTC m=+151.340414815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.039610 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.039943 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.077502 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.113141 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.113203 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.113226 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" event={"ID":"020fd783-0219-4f9d-8504-f430877d9ed4","Type":"ContainerStarted","Data":"ed1731b46423f3de76a4c3453d45680e5523482b34fe5ab61c9c412fd9aa3ec0"} Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.120038 4696 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wkn7w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.120089 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.130628 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.130924 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.630875836 +0000 UTC m=+151.444492423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.131236 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.131346 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.131454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.133889 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.633877549 +0000 UTC m=+151.447494136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.139074 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nd9j4"] Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.232634 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.233003 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.233050 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.236702 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.236886 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.736823762 +0000 UTC m=+151.550440359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.239164 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.242777 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.742757765 +0000 UTC m=+151.556374422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.282824 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.344453 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.345186 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.845169303 +0000 UTC m=+151.658785890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.360985 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.410803 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:14 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:14 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:14 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.410848 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.451402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.462334 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:14.962312494 +0000 UTC m=+151.775929071 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.517690 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.589461 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.589925 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.089898037 +0000 UTC m=+151.903514624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.693402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.694451 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.19443876 +0000 UTC m=+152.008055347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.794600 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.793892 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tpmx6"] Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.794879 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.294852136 +0000 UTC m=+152.108468723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.795081 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.795366 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.295359651 +0000 UTC m=+152.108976238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:14 crc kubenswrapper[4696]: I1125 10:39:14.902219 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:14 crc kubenswrapper[4696]: E1125 10:39:14.902771 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.402749321 +0000 UTC m=+152.216365908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.023348 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.023634 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.523622708 +0000 UTC m=+152.337239295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.048293 4696 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-2pqps container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.048350 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" podUID="a0679b82-8f22-4227-bdbf-d92e063ffba0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.048924 4696 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-2pqps container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.048945 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" podUID="a0679b82-8f22-4227-bdbf-d92e063ffba0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.091578 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.091641 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.092003 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.092018 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.115603 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-szcdp" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.126593 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.127199 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.62718372 +0000 UTC m=+152.440800307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.160166 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" event={"ID":"020fd783-0219-4f9d-8504-f430877d9ed4","Type":"ContainerStarted","Data":"ed1206c06891e59de43a7ec7bc12f1e79bbb3bdd775d76d4b76a092e4e10cafd"} Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.165009 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-88drx"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.165563 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerID="7fff65983db5a1d5a438cd3f871f6aab6e1ee0f1fc63eb4495e4cd55ebd19925" exitCode=0 Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.165602 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd9j4" event={"ID":"d6b24f5f-89c0-4078-be5d-8739b554e76e","Type":"ContainerDied","Data":"7fff65983db5a1d5a438cd3f871f6aab6e1ee0f1fc63eb4495e4cd55ebd19925"} Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.165617 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd9j4" event={"ID":"d6b24f5f-89c0-4078-be5d-8739b554e76e","Type":"ContainerStarted","Data":"a916926c82d28802b4e4284e74165c4a61522692744fe42c7ce65c8d315005b2"} Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.184696 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.209065 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"761a22b39ada5f3a59f06047054979fea5965f34cdd8ff936053b432a7b28a62"} Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.209678 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.228740 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.230086 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.730069982 +0000 UTC m=+152.543686569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.257703 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8kvxk"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.258682 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.259017 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"52000db1bc507401807299df89bb144d42d88c1fb85540e9ed86a62d3ae3a115"} Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.261748 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.280286 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpmx6" event={"ID":"46840e91-cc38-4a5c-beca-96f3735590c4","Type":"ContainerStarted","Data":"891ecc107ecb77c4c750eba7266c0cc140437b8dd700b0fb74d497919d5dd82f"} Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.332604 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.332905 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.832880482 +0000 UTC m=+152.646497069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.333562 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.333676 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-catalog-content\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.334873 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-utilities\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.335049 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q99qr\" (UniqueName: \"kubernetes.io/projected/1150bbc8-8968-47f9-be96-81e1de5b8043-kube-api-access-q99qr\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.335239 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.835229384 +0000 UTC m=+152.648845971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.359053 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kvxk"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.390307 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.405145 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:15 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:15 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:15 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.405213 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.438637 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.439823 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.440947 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.443759 4696 patch_prober.go:28] interesting pod/console-f9d7485db-rj85t container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.443952 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rj85t" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.454730 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.455137 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-catalog-content\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.455234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-utilities\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.455287 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q99qr\" (UniqueName: \"kubernetes.io/projected/1150bbc8-8968-47f9-be96-81e1de5b8043-kube-api-access-q99qr\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.456328 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:15.956295557 +0000 UTC m=+152.769912144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.457248 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-utilities\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.457589 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-catalog-content\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.485156 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.485203 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.566041 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.594761 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.094724694 +0000 UTC m=+152.908341281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.615577 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvmp4"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.645072 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q99qr\" (UniqueName: \"kubernetes.io/projected/1150bbc8-8968-47f9-be96-81e1de5b8043-kube-api-access-q99qr\") pod \"redhat-marketplace-8kvxk\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.657570 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8fmqn"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.664103 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.666634 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8fmqn"] Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.670480 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.670925 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.170908132 +0000 UTC m=+152.984524719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.771580 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-catalog-content\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.771625 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-utilities\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.771648 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxdss\" (UniqueName: \"kubernetes.io/projected/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-kube-api-access-xxdss\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.771728 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.772021 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.272007009 +0000 UTC m=+153.085623596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.822621 4696 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wkn7w container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.822990 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.822751 4696 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wkn7w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.823122 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.879280 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.879458 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-catalog-content\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.879486 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-utilities\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.879508 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxdss\" (UniqueName: \"kubernetes.io/projected/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-kube-api-access-xxdss\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.879864 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.379838463 +0000 UTC m=+153.193455050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.880202 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-utilities\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.880302 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-catalog-content\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.901438 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxdss\" (UniqueName: \"kubernetes.io/projected/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-kube-api-access-xxdss\") pod \"redhat-marketplace-8fmqn\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.917364 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:15 crc kubenswrapper[4696]: I1125 10:39:15.985551 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:15 crc kubenswrapper[4696]: E1125 10:39:15.985949 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.485936844 +0000 UTC m=+153.299553431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.006972 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.087064 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:16 crc kubenswrapper[4696]: E1125 10:39:16.087492 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.587474164 +0000 UTC m=+153.401090751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.180298 4696 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.188350 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:16 crc kubenswrapper[4696]: E1125 10:39:16.188790 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:39:16.688778397 +0000 UTC m=+153.502394984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xkw2f" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.220775 4696 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T10:39:16.180553094Z","Handler":null,"Name":""} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.228948 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-52nv8"] Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.230247 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.233496 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.235221 4696 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.235260 4696 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.249743 4696 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bv2wn container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]log ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]etcd ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/max-in-flight-filter ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 25 10:39:16 crc kubenswrapper[4696]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 25 10:39:16 crc kubenswrapper[4696]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/project.openshift.io-projectcache ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-startinformers ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 25 10:39:16 crc kubenswrapper[4696]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 10:39:16 crc kubenswrapper[4696]: livez check failed Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.249799 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" podUID="29a2d684-8d98-42c3-9d91-12d0a87bf032" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.289199 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.289452 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-catalog-content\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.289499 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k58vp\" (UniqueName: \"kubernetes.io/projected/2d0338a7-bada-47ba-a1b9-21fd05a46f55-kube-api-access-k58vp\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.289518 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-utilities\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.298214 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52nv8"] Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.315518 4696 generic.go:334] "Generic (PLEG): container finished" podID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerID="834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd" exitCode=0 Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.315617 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerDied","Data":"834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.315646 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerStarted","Data":"62ae701cac36ff8537e4ff13c02158ae2a333fe137040ee3e8495cdc6b2bac4c"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.318383 4696 generic.go:334] "Generic (PLEG): container finished" podID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerID="39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25" exitCode=0 Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.318483 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88drx" event={"ID":"a42c7819-a52b-4dbc-bbe5-528c00fb6d95","Type":"ContainerDied","Data":"39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.318514 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88drx" event={"ID":"a42c7819-a52b-4dbc-bbe5-528c00fb6d95","Type":"ContainerStarted","Data":"8f35ab4fc84000675215ca50c5526171b08a605550b7737889065b784194ab6f"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.349160 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.349653 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" event={"ID":"020fd783-0219-4f9d-8504-f430877d9ed4","Type":"ContainerStarted","Data":"2170a367fe4a2bbda76cb5134be0a06be86d6a9cb15af1cb370b53b6130a011b"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.375181 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4bb6022e-d54f-4374-bc8b-4dd422cd0a36","Type":"ContainerStarted","Data":"b7b84080ccbccb4d64b3860bd7aab09aa7133a44623655ac43f718d6d501436c"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.375222 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4bb6022e-d54f-4374-bc8b-4dd422cd0a36","Type":"ContainerStarted","Data":"1d7d55e13144e54a97e99a0a657ff4904d45ab61c7a6e659e7c5fe411cf11f48"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.391335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-utilities\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.391396 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.391509 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-catalog-content\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.391565 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k58vp\" (UniqueName: \"kubernetes.io/projected/2d0338a7-bada-47ba-a1b9-21fd05a46f55-kube-api-access-k58vp\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.391937 4696 generic.go:334] "Generic (PLEG): container finished" podID="46840e91-cc38-4a5c-beca-96f3735590c4" containerID="ddb0e1db24961ed8b91e5e64177cf72255e6668b196a9eb15f35edfd24019be6" exitCode=0 Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.392970 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpmx6" event={"ID":"46840e91-cc38-4a5c-beca-96f3735590c4","Type":"ContainerDied","Data":"ddb0e1db24961ed8b91e5e64177cf72255e6668b196a9eb15f35edfd24019be6"} Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.393375 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-utilities\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.394720 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-catalog-content\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.425954 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:16 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:16 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:16 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.426267 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.435605 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k58vp\" (UniqueName: \"kubernetes.io/projected/2d0338a7-bada-47ba-a1b9-21fd05a46f55-kube-api-access-k58vp\") pod \"redhat-operators-52nv8\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.443218 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.443257 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.554750 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.567994 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" podStartSLOduration=14.567975048 podStartE2EDuration="14.567975048s" podCreationTimestamp="2025-11-25 10:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:16.536419674 +0000 UTC m=+153.350036261" watchObservedRunningTime="2025-11-25 10:39:16.567975048 +0000 UTC m=+153.381591635" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.657804 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.657786216 podStartE2EDuration="2.657786216s" podCreationTimestamp="2025-11-25 10:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:16.606548747 +0000 UTC m=+153.420165334" watchObservedRunningTime="2025-11-25 10:39:16.657786216 +0000 UTC m=+153.471402803" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.658647 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j6j5q"] Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.659600 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.688287 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xkw2f\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.688898 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j6j5q"] Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.710055 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55gj2\" (UniqueName: \"kubernetes.io/projected/1b409cdd-df72-4365-983d-852916369a5a-kube-api-access-55gj2\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.710129 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-catalog-content\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.710174 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-utilities\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.771999 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.812506 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-catalog-content\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.813094 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-utilities\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.813919 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55gj2\" (UniqueName: \"kubernetes.io/projected/1b409cdd-df72-4365-983d-852916369a5a-kube-api-access-55gj2\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.815080 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-catalog-content\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.815366 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-utilities\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.831037 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kvxk"] Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.845831 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55gj2\" (UniqueName: \"kubernetes.io/projected/1b409cdd-df72-4365-983d-852916369a5a-kube-api-access-55gj2\") pod \"redhat-operators-j6j5q\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:16 crc kubenswrapper[4696]: I1125 10:39:16.996206 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.009583 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8fmqn"] Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.048855 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52nv8"] Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.058204 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2pqps" Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.399985 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:17 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:17 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:17 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.400026 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.404744 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerStarted","Data":"193af7cd19f9e48736658baff6c263402e220802dfcf6c03cc8fae9e526e8505"} Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.408472 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8fmqn" event={"ID":"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01","Type":"ContainerStarted","Data":"4b96d5f4ad26900decf934da743d5c90c205fddfdde2439cb3caa689a275dac2"} Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.419997 4696 generic.go:334] "Generic (PLEG): container finished" podID="0e6edd3c-d469-4a55-a9b8-2a3087248db6" containerID="9881f9e2a38942d186ed0835edac34e728e29c17093199894b1b465a1db89cc8" exitCode=0 Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.420072 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" event={"ID":"0e6edd3c-d469-4a55-a9b8-2a3087248db6","Type":"ContainerDied","Data":"9881f9e2a38942d186ed0835edac34e728e29c17093199894b1b465a1db89cc8"} Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.425254 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j6j5q"] Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.426768 4696 generic.go:334] "Generic (PLEG): container finished" podID="4bb6022e-d54f-4374-bc8b-4dd422cd0a36" containerID="b7b84080ccbccb4d64b3860bd7aab09aa7133a44623655ac43f718d6d501436c" exitCode=0 Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.426821 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4bb6022e-d54f-4374-bc8b-4dd422cd0a36","Type":"ContainerDied","Data":"b7b84080ccbccb4d64b3860bd7aab09aa7133a44623655ac43f718d6d501436c"} Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.443370 4696 generic.go:334] "Generic (PLEG): container finished" podID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerID="06e7ab4b39b6fbe4622bc213bcca677b6a6e6938ea5236eb7c0e80f647e9e1a8" exitCode=0 Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.443727 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kvxk" event={"ID":"1150bbc8-8968-47f9-be96-81e1de5b8043","Type":"ContainerDied","Data":"06e7ab4b39b6fbe4622bc213bcca677b6a6e6938ea5236eb7c0e80f647e9e1a8"} Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.443774 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kvxk" event={"ID":"1150bbc8-8968-47f9-be96-81e1de5b8043","Type":"ContainerStarted","Data":"1a600f4d43912673573a95cfd509d9842c2ff8fcfd26c117802a8003fac8e6cc"} Nov 25 10:39:17 crc kubenswrapper[4696]: I1125 10:39:17.477308 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xkw2f"] Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.052635 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.390851 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:18 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:18 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:18 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.391104 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.451145 4696 generic.go:334] "Generic (PLEG): container finished" podID="1b409cdd-df72-4365-983d-852916369a5a" containerID="2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d" exitCode=0 Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.451241 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6j5q" event={"ID":"1b409cdd-df72-4365-983d-852916369a5a","Type":"ContainerDied","Data":"2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d"} Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.451277 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6j5q" event={"ID":"1b409cdd-df72-4365-983d-852916369a5a","Type":"ContainerStarted","Data":"6d392815d876c681756e1572fa45520540a5bfc331c5de60315ed094538d0a3d"} Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.458554 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" event={"ID":"545452b4-d392-4d46-8e5e-a2a4c9e7b43b","Type":"ContainerStarted","Data":"19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362"} Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.458617 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" event={"ID":"545452b4-d392-4d46-8e5e-a2a4c9e7b43b","Type":"ContainerStarted","Data":"349ba1afc3174da7195e9fe5015cf8c45a83e90387e12f396583a89fe0a90f77"} Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.458708 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.470993 4696 generic.go:334] "Generic (PLEG): container finished" podID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerID="b291ae23dab9bd3d9239d7b9bd9b2ffd2b170e08b8d7303ae58ba572f9a19c3d" exitCode=0 Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.471139 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerDied","Data":"b291ae23dab9bd3d9239d7b9bd9b2ffd2b170e08b8d7303ae58ba572f9a19c3d"} Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.490061 4696 generic.go:334] "Generic (PLEG): container finished" podID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerID="56f381eea2d3fd1652645b7e196531e080c388a8e3cf019ed99597e233b781d5" exitCode=0 Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.490249 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8fmqn" event={"ID":"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01","Type":"ContainerDied","Data":"56f381eea2d3fd1652645b7e196531e080c388a8e3cf019ed99597e233b781d5"} Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.503473 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" podStartSLOduration=134.503448753 podStartE2EDuration="2m14.503448753s" podCreationTimestamp="2025-11-25 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:18.497416587 +0000 UTC m=+155.311033194" watchObservedRunningTime="2025-11-25 10:39:18.503448753 +0000 UTC m=+155.317065340" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.794562 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.805097 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.813184 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.813499 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.824049 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.860178 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e55e2607-be7d-4211-9b08-cf6905029542-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.860235 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e55e2607-be7d-4211-9b08-cf6905029542-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.961449 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e55e2607-be7d-4211-9b08-cf6905029542-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.961497 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e55e2607-be7d-4211-9b08-cf6905029542-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:18 crc kubenswrapper[4696]: I1125 10:39:18.961614 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e55e2607-be7d-4211-9b08-cf6905029542-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.020440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e55e2607-be7d-4211-9b08-cf6905029542-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.065054 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.065797 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.140032 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.166410 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e6edd3c-d469-4a55-a9b8-2a3087248db6-config-volume\") pod \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.166468 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kube-api-access\") pod \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.166653 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kubelet-dir\") pod \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\" (UID: \"4bb6022e-d54f-4374-bc8b-4dd422cd0a36\") " Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.166734 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e6edd3c-d469-4a55-a9b8-2a3087248db6-secret-volume\") pod \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.166781 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jll8\" (UniqueName: \"kubernetes.io/projected/0e6edd3c-d469-4a55-a9b8-2a3087248db6-kube-api-access-7jll8\") pod \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\" (UID: \"0e6edd3c-d469-4a55-a9b8-2a3087248db6\") " Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.169302 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e6edd3c-d469-4a55-a9b8-2a3087248db6-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e6edd3c-d469-4a55-a9b8-2a3087248db6" (UID: "0e6edd3c-d469-4a55-a9b8-2a3087248db6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.169387 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4bb6022e-d54f-4374-bc8b-4dd422cd0a36" (UID: "4bb6022e-d54f-4374-bc8b-4dd422cd0a36"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.180484 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6edd3c-d469-4a55-a9b8-2a3087248db6-kube-api-access-7jll8" (OuterVolumeSpecName: "kube-api-access-7jll8") pod "0e6edd3c-d469-4a55-a9b8-2a3087248db6" (UID: "0e6edd3c-d469-4a55-a9b8-2a3087248db6"). InnerVolumeSpecName "kube-api-access-7jll8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.181589 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e6edd3c-d469-4a55-a9b8-2a3087248db6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e6edd3c-d469-4a55-a9b8-2a3087248db6" (UID: "0e6edd3c-d469-4a55-a9b8-2a3087248db6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.188010 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4bb6022e-d54f-4374-bc8b-4dd422cd0a36" (UID: "4bb6022e-d54f-4374-bc8b-4dd422cd0a36"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.269072 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.269106 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e6edd3c-d469-4a55-a9b8-2a3087248db6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.269116 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4bb6022e-d54f-4374-bc8b-4dd422cd0a36-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.269126 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e6edd3c-d469-4a55-a9b8-2a3087248db6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.269135 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jll8\" (UniqueName: \"kubernetes.io/projected/0e6edd3c-d469-4a55-a9b8-2a3087248db6-kube-api-access-7jll8\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.394074 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:19 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:19 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:19 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.394186 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.521470 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" event={"ID":"0e6edd3c-d469-4a55-a9b8-2a3087248db6","Type":"ContainerDied","Data":"bc7cac1de148f4c3688cf6a1bf7bda623c61e2347f3554d52704f1ac383b2ad3"} Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.521505 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc7cac1de148f4c3688cf6a1bf7bda623c61e2347f3554d52704f1ac383b2ad3" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.521563 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.527379 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.527340 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4bb6022e-d54f-4374-bc8b-4dd422cd0a36","Type":"ContainerDied","Data":"1d7d55e13144e54a97e99a0a657ff4904d45ab61c7a6e659e7c5fe411cf11f48"} Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.527442 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d7d55e13144e54a97e99a0a657ff4904d45ab61c7a6e659e7c5fe411cf11f48" Nov 25 10:39:19 crc kubenswrapper[4696]: I1125 10:39:19.813090 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 10:39:20 crc kubenswrapper[4696]: I1125 10:39:20.393791 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:20 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:20 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:20 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:20 crc kubenswrapper[4696]: I1125 10:39:20.394228 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:20 crc kubenswrapper[4696]: I1125 10:39:20.487388 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:20 crc kubenswrapper[4696]: I1125 10:39:20.496653 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-bv2wn" Nov 25 10:39:20 crc kubenswrapper[4696]: I1125 10:39:20.615104 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e55e2607-be7d-4211-9b08-cf6905029542","Type":"ContainerStarted","Data":"570111c31f134c402b807af7bcc5de64c75812f772c5f2d3e4474c89d5c0f75f"} Nov 25 10:39:20 crc kubenswrapper[4696]: I1125 10:39:20.875595 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dkss6" Nov 25 10:39:21 crc kubenswrapper[4696]: I1125 10:39:21.392319 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:21 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:21 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:21 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:21 crc kubenswrapper[4696]: I1125 10:39:21.392376 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:21 crc kubenswrapper[4696]: I1125 10:39:21.660931 4696 generic.go:334] "Generic (PLEG): container finished" podID="e55e2607-be7d-4211-9b08-cf6905029542" containerID="c5309affc059d1a0dee0414cb9a38306d6fcc0b4baac6ccc72f40a76be4e5878" exitCode=0 Nov 25 10:39:21 crc kubenswrapper[4696]: I1125 10:39:21.660972 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e55e2607-be7d-4211-9b08-cf6905029542","Type":"ContainerDied","Data":"c5309affc059d1a0dee0414cb9a38306d6fcc0b4baac6ccc72f40a76be4e5878"} Nov 25 10:39:22 crc kubenswrapper[4696]: I1125 10:39:22.393058 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:22 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:22 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:22 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:22 crc kubenswrapper[4696]: I1125 10:39:22.393132 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.184474 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.270241 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e55e2607-be7d-4211-9b08-cf6905029542-kube-api-access\") pod \"e55e2607-be7d-4211-9b08-cf6905029542\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.270282 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e55e2607-be7d-4211-9b08-cf6905029542-kubelet-dir\") pod \"e55e2607-be7d-4211-9b08-cf6905029542\" (UID: \"e55e2607-be7d-4211-9b08-cf6905029542\") " Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.270418 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e55e2607-be7d-4211-9b08-cf6905029542-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e55e2607-be7d-4211-9b08-cf6905029542" (UID: "e55e2607-be7d-4211-9b08-cf6905029542"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.270623 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e55e2607-be7d-4211-9b08-cf6905029542-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.275734 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e55e2607-be7d-4211-9b08-cf6905029542-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e55e2607-be7d-4211-9b08-cf6905029542" (UID: "e55e2607-be7d-4211-9b08-cf6905029542"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.371635 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e55e2607-be7d-4211-9b08-cf6905029542-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.391251 4696 patch_prober.go:28] interesting pod/router-default-5444994796-k24kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:39:23 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Nov 25 10:39:23 crc kubenswrapper[4696]: [+]process-running ok Nov 25 10:39:23 crc kubenswrapper[4696]: healthz check failed Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.391316 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k24kz" podUID="84bba043-a192-4ea4-9225-73cc33b31ea5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.704340 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e55e2607-be7d-4211-9b08-cf6905029542","Type":"ContainerDied","Data":"570111c31f134c402b807af7bcc5de64c75812f772c5f2d3e4474c89d5c0f75f"} Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.704377 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="570111c31f134c402b807af7bcc5de64c75812f772c5f2d3e4474c89d5c0f75f" Nov 25 10:39:23 crc kubenswrapper[4696]: I1125 10:39:23.704431 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:39:24 crc kubenswrapper[4696]: I1125 10:39:24.391964 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:24 crc kubenswrapper[4696]: I1125 10:39:24.396395 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-k24kz" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.091453 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.091555 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.091576 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-mjb8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.091695 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mjb8l" podUID="e6d2a426-4cba-4b45-bce3-45633b970e8e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.219015 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.246328 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19727774-75c0-47f9-9157-75b1116d7daa-metrics-certs\") pod \"network-metrics-daemon-w7l9g\" (UID: \"19727774-75c0-47f9-9157-75b1116d7daa\") " pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.436415 4696 patch_prober.go:28] interesting pod/console-f9d7485db-rj85t container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.436651 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rj85t" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.468925 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w7l9g" Nov 25 10:39:25 crc kubenswrapper[4696]: I1125 10:39:25.823152 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:39:30 crc kubenswrapper[4696]: I1125 10:39:30.801535 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:39:30 crc kubenswrapper[4696]: I1125 10:39:30.801877 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:39:35 crc kubenswrapper[4696]: I1125 10:39:35.096682 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mjb8l" Nov 25 10:39:35 crc kubenswrapper[4696]: I1125 10:39:35.443430 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:35 crc kubenswrapper[4696]: I1125 10:39:35.448476 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:39:36 crc kubenswrapper[4696]: I1125 10:39:36.785978 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:39:45 crc kubenswrapper[4696]: I1125 10:39:45.491890 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wzm5m" Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.452609 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-w7l9g"] Nov 25 10:39:47 crc kubenswrapper[4696]: W1125 10:39:47.468927 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19727774_75c0_47f9_9157_75b1116d7daa.slice/crio-9886a54b4f5f5c598656864461557e717f8adb39f641fa91af4e8026807ec80d WatchSource:0}: Error finding container 9886a54b4f5f5c598656864461557e717f8adb39f641fa91af4e8026807ec80d: Status 404 returned error can't find the container with id 9886a54b4f5f5c598656864461557e717f8adb39f641fa91af4e8026807ec80d Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.883708 4696 generic.go:334] "Generic (PLEG): container finished" podID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerID="c35578be44aed651f8504eeca8c0fe142b087b9b6f2bf70be5e7244bd48ab8d7" exitCode=0 Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.883798 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8fmqn" event={"ID":"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01","Type":"ContainerDied","Data":"c35578be44aed651f8504eeca8c0fe142b087b9b6f2bf70be5e7244bd48ab8d7"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.884899 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" event={"ID":"19727774-75c0-47f9-9157-75b1116d7daa","Type":"ContainerStarted","Data":"9886a54b4f5f5c598656864461557e717f8adb39f641fa91af4e8026807ec80d"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.887855 4696 generic.go:334] "Generic (PLEG): container finished" podID="46840e91-cc38-4a5c-beca-96f3735590c4" containerID="76ed17a3f8550e38520fc69ccb4edaa54e050e2c41d0db6ef023e163a870a423" exitCode=0 Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.887937 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpmx6" event={"ID":"46840e91-cc38-4a5c-beca-96f3735590c4","Type":"ContainerDied","Data":"76ed17a3f8550e38520fc69ccb4edaa54e050e2c41d0db6ef023e163a870a423"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.890045 4696 generic.go:334] "Generic (PLEG): container finished" podID="1b409cdd-df72-4365-983d-852916369a5a" containerID="0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14" exitCode=0 Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.890095 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6j5q" event={"ID":"1b409cdd-df72-4365-983d-852916369a5a","Type":"ContainerDied","Data":"0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.898516 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerStarted","Data":"0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.903689 4696 generic.go:334] "Generic (PLEG): container finished" podID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerID="83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1" exitCode=0 Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.903785 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88drx" event={"ID":"a42c7819-a52b-4dbc-bbe5-528c00fb6d95","Type":"ContainerDied","Data":"83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.909321 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerStarted","Data":"f4e30fe4862c4d4e2cc769b34ae0138d2513a4203e0f8e80afe8fb031de04918"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.911414 4696 generic.go:334] "Generic (PLEG): container finished" podID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerID="5f924c3562f5458f055a6456f870ac20e06d5a928d0a22f1d497411c08e0ce26" exitCode=0 Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.911499 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kvxk" event={"ID":"1150bbc8-8968-47f9-be96-81e1de5b8043","Type":"ContainerDied","Data":"5f924c3562f5458f055a6456f870ac20e06d5a928d0a22f1d497411c08e0ce26"} Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.923858 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerID="47e5459d1a409d36eca96df2186bada0a84143b5724a0f67820b244f81bd6429" exitCode=0 Nov 25 10:39:47 crc kubenswrapper[4696]: I1125 10:39:47.923900 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd9j4" event={"ID":"d6b24f5f-89c0-4078-be5d-8739b554e76e","Type":"ContainerDied","Data":"47e5459d1a409d36eca96df2186bada0a84143b5724a0f67820b244f81bd6429"} Nov 25 10:39:48 crc kubenswrapper[4696]: I1125 10:39:48.945908 4696 generic.go:334] "Generic (PLEG): container finished" podID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerID="0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f" exitCode=0 Nov 25 10:39:48 crc kubenswrapper[4696]: I1125 10:39:48.946011 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerDied","Data":"0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f"} Nov 25 10:39:48 crc kubenswrapper[4696]: I1125 10:39:48.950927 4696 generic.go:334] "Generic (PLEG): container finished" podID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerID="f4e30fe4862c4d4e2cc769b34ae0138d2513a4203e0f8e80afe8fb031de04918" exitCode=0 Nov 25 10:39:48 crc kubenswrapper[4696]: I1125 10:39:48.950994 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerDied","Data":"f4e30fe4862c4d4e2cc769b34ae0138d2513a4203e0f8e80afe8fb031de04918"} Nov 25 10:39:48 crc kubenswrapper[4696]: I1125 10:39:48.958040 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" event={"ID":"19727774-75c0-47f9-9157-75b1116d7daa","Type":"ContainerStarted","Data":"1c6dfe64708bfe6ebe053910661d11e3d7693b098707f3f8ee793135d49ff83b"} Nov 25 10:39:49 crc kubenswrapper[4696]: I1125 10:39:49.966961 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-w7l9g" event={"ID":"19727774-75c0-47f9-9157-75b1116d7daa","Type":"ContainerStarted","Data":"214d46c6080c3c0049d29b3f287a572f18efcd0e5daf617bd18de14c03883354"} Nov 25 10:39:49 crc kubenswrapper[4696]: I1125 10:39:49.993930 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-w7l9g" podStartSLOduration=166.993813772 podStartE2EDuration="2m46.993813772s" podCreationTimestamp="2025-11-25 10:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:49.985272479 +0000 UTC m=+186.798889106" watchObservedRunningTime="2025-11-25 10:39:49.993813772 +0000 UTC m=+186.807430399" Nov 25 10:39:50 crc kubenswrapper[4696]: I1125 10:39:50.978349 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6j5q" event={"ID":"1b409cdd-df72-4365-983d-852916369a5a","Type":"ContainerStarted","Data":"523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6"} Nov 25 10:39:51 crc kubenswrapper[4696]: I1125 10:39:51.990724 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:39:52 crc kubenswrapper[4696]: I1125 10:39:52.012322 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j6j5q" podStartSLOduration=4.026582118 podStartE2EDuration="36.012300918s" podCreationTimestamp="2025-11-25 10:39:16 +0000 UTC" firstStartedPulling="2025-11-25 10:39:18.454085792 +0000 UTC m=+155.267702379" lastFinishedPulling="2025-11-25 10:39:50.439804582 +0000 UTC m=+187.253421179" observedRunningTime="2025-11-25 10:39:51.010326119 +0000 UTC m=+187.823942726" watchObservedRunningTime="2025-11-25 10:39:52.012300918 +0000 UTC m=+188.825917515" Nov 25 10:39:52 crc kubenswrapper[4696]: I1125 10:39:52.993473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8fmqn" event={"ID":"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01","Type":"ContainerStarted","Data":"e5acb745c267495a4abed05e8e8b56ef598db11ca66814416f76085992390b8d"} Nov 25 10:39:52 crc kubenswrapper[4696]: I1125 10:39:52.995901 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerStarted","Data":"4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72"} Nov 25 10:39:53 crc kubenswrapper[4696]: I1125 10:39:53.015988 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8fmqn" podStartSLOduration=4.894704072 podStartE2EDuration="38.015968248s" podCreationTimestamp="2025-11-25 10:39:15 +0000 UTC" firstStartedPulling="2025-11-25 10:39:18.492323 +0000 UTC m=+155.305939587" lastFinishedPulling="2025-11-25 10:39:51.613587176 +0000 UTC m=+188.427203763" observedRunningTime="2025-11-25 10:39:53.015893496 +0000 UTC m=+189.829510083" watchObservedRunningTime="2025-11-25 10:39:53.015968248 +0000 UTC m=+189.829584835" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.003755 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88drx" event={"ID":"a42c7819-a52b-4dbc-bbe5-528c00fb6d95","Type":"ContainerStarted","Data":"78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9"} Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.005680 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kvxk" event={"ID":"1150bbc8-8968-47f9-be96-81e1de5b8043","Type":"ContainerStarted","Data":"f3d5ec472d0f2280f4323178a595cebde0cd524ff043f0477b66801ffd4b8c0d"} Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.007640 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd9j4" event={"ID":"d6b24f5f-89c0-4078-be5d-8739b554e76e","Type":"ContainerStarted","Data":"52e56e9d94d3d4d5a333fcf1797220f56c4562b119f5a797de0b6df0d1732c24"} Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.009539 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerStarted","Data":"4bb5219ac9ca64c7654c2102a570cc66aa7d2df7e73eee7431d2fdf2d8ebc5f9"} Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.011249 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpmx6" event={"ID":"46840e91-cc38-4a5c-beca-96f3735590c4","Type":"ContainerStarted","Data":"88e90f3f438b54c08800951e7cd9fe24176256260937938c2e2663352983cbcb"} Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.023787 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-88drx" podStartSLOduration=4.221631198 podStartE2EDuration="41.023768857s" podCreationTimestamp="2025-11-25 10:39:13 +0000 UTC" firstStartedPulling="2025-11-25 10:39:16.336861752 +0000 UTC m=+153.150478339" lastFinishedPulling="2025-11-25 10:39:53.138999421 +0000 UTC m=+189.952615998" observedRunningTime="2025-11-25 10:39:54.022788767 +0000 UTC m=+190.836405374" watchObservedRunningTime="2025-11-25 10:39:54.023768857 +0000 UTC m=+190.837385444" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.050408 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8kvxk" podStartSLOduration=3.423800359 podStartE2EDuration="39.050387308s" podCreationTimestamp="2025-11-25 10:39:15 +0000 UTC" firstStartedPulling="2025-11-25 10:39:17.448955066 +0000 UTC m=+154.262571643" lastFinishedPulling="2025-11-25 10:39:53.075542005 +0000 UTC m=+189.889158592" observedRunningTime="2025-11-25 10:39:54.049894492 +0000 UTC m=+190.863511099" watchObservedRunningTime="2025-11-25 10:39:54.050387308 +0000 UTC m=+190.864003895" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.079074 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.079387 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.107705 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-52nv8" podStartSLOduration=3.384091352 podStartE2EDuration="38.107655193s" podCreationTimestamp="2025-11-25 10:39:16 +0000 UTC" firstStartedPulling="2025-11-25 10:39:18.480379332 +0000 UTC m=+155.293995919" lastFinishedPulling="2025-11-25 10:39:53.203943183 +0000 UTC m=+190.017559760" observedRunningTime="2025-11-25 10:39:54.076959216 +0000 UTC m=+190.890575803" watchObservedRunningTime="2025-11-25 10:39:54.107655193 +0000 UTC m=+190.921271780" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.132570 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nd9j4" podStartSLOduration=3.6888741940000003 podStartE2EDuration="41.13253618s" podCreationTimestamp="2025-11-25 10:39:13 +0000 UTC" firstStartedPulling="2025-11-25 10:39:15.184418005 +0000 UTC m=+151.998034592" lastFinishedPulling="2025-11-25 10:39:52.628080001 +0000 UTC m=+189.441696578" observedRunningTime="2025-11-25 10:39:54.112210874 +0000 UTC m=+190.925827471" watchObservedRunningTime="2025-11-25 10:39:54.13253618 +0000 UTC m=+190.946152767" Nov 25 10:39:54 crc kubenswrapper[4696]: I1125 10:39:54.157525 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tpmx6" podStartSLOduration=4.915055195 podStartE2EDuration="41.15750753s" podCreationTimestamp="2025-11-25 10:39:13 +0000 UTC" firstStartedPulling="2025-11-25 10:39:16.396038907 +0000 UTC m=+153.209655494" lastFinishedPulling="2025-11-25 10:39:52.638491252 +0000 UTC m=+189.452107829" observedRunningTime="2025-11-25 10:39:54.137643678 +0000 UTC m=+190.951260265" watchObservedRunningTime="2025-11-25 10:39:54.15750753 +0000 UTC m=+190.971124117" Nov 25 10:39:55 crc kubenswrapper[4696]: I1125 10:39:55.313683 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jvmp4" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="registry-server" probeResult="failure" output=< Nov 25 10:39:55 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 10:39:55 crc kubenswrapper[4696]: > Nov 25 10:39:55 crc kubenswrapper[4696]: I1125 10:39:55.918152 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:55 crc kubenswrapper[4696]: I1125 10:39:55.918396 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:55 crc kubenswrapper[4696]: I1125 10:39:55.961978 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:39:55 crc kubenswrapper[4696]: I1125 10:39:55.977596 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jvmp4" podStartSLOduration=6.6897802760000005 podStartE2EDuration="42.977580789s" podCreationTimestamp="2025-11-25 10:39:13 +0000 UTC" firstStartedPulling="2025-11-25 10:39:16.317205676 +0000 UTC m=+153.130822263" lastFinishedPulling="2025-11-25 10:39:52.605006189 +0000 UTC m=+189.418622776" observedRunningTime="2025-11-25 10:39:54.162127542 +0000 UTC m=+190.975744129" watchObservedRunningTime="2025-11-25 10:39:55.977580789 +0000 UTC m=+192.791197366" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.007692 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.007735 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.049709 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.556061 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.556101 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.997159 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:56 crc kubenswrapper[4696]: I1125 10:39:56.997523 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:39:57 crc kubenswrapper[4696]: I1125 10:39:57.596510 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-52nv8" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="registry-server" probeResult="failure" output=< Nov 25 10:39:57 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 10:39:57 crc kubenswrapper[4696]: > Nov 25 10:39:58 crc kubenswrapper[4696]: I1125 10:39:58.048010 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j6j5q" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="registry-server" probeResult="failure" output=< Nov 25 10:39:58 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 10:39:58 crc kubenswrapper[4696]: > Nov 25 10:40:00 crc kubenswrapper[4696]: I1125 10:40:00.801637 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:40:00 crc kubenswrapper[4696]: I1125 10:40:00.801727 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.387553 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.387824 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.434802 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.594695 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.594991 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.630038 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.849555 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.849613 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:40:03 crc kubenswrapper[4696]: I1125 10:40:03.887907 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:40:04 crc kubenswrapper[4696]: I1125 10:40:04.085430 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nvzm5"] Nov 25 10:40:04 crc kubenswrapper[4696]: I1125 10:40:04.136977 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:40:04 crc kubenswrapper[4696]: I1125 10:40:04.150601 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:40:04 crc kubenswrapper[4696]: I1125 10:40:04.153061 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:40:04 crc kubenswrapper[4696]: I1125 10:40:04.174287 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:40:04 crc kubenswrapper[4696]: I1125 10:40:04.228084 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:40:05 crc kubenswrapper[4696]: I1125 10:40:05.064184 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-88drx"] Nov 25 10:40:05 crc kubenswrapper[4696]: I1125 10:40:05.962913 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.050810 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.072398 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-88drx" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="registry-server" containerID="cri-o://78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9" gracePeriod=2 Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.416001 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.464469 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvmp4"] Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.464980 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jvmp4" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="registry-server" containerID="cri-o://4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72" gracePeriod=2 Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.554387 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-utilities\") pod \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.554516 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-catalog-content\") pod \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.554568 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v942r\" (UniqueName: \"kubernetes.io/projected/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-kube-api-access-v942r\") pod \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\" (UID: \"a42c7819-a52b-4dbc-bbe5-528c00fb6d95\") " Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.555162 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-utilities" (OuterVolumeSpecName: "utilities") pod "a42c7819-a52b-4dbc-bbe5-528c00fb6d95" (UID: "a42c7819-a52b-4dbc-bbe5-528c00fb6d95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.559832 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-kube-api-access-v942r" (OuterVolumeSpecName: "kube-api-access-v942r") pod "a42c7819-a52b-4dbc-bbe5-528c00fb6d95" (UID: "a42c7819-a52b-4dbc-bbe5-528c00fb6d95"). InnerVolumeSpecName "kube-api-access-v942r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.593906 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.602002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a42c7819-a52b-4dbc-bbe5-528c00fb6d95" (UID: "a42c7819-a52b-4dbc-bbe5-528c00fb6d95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.636923 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.656356 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.656396 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v942r\" (UniqueName: \"kubernetes.io/projected/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-kube-api-access-v942r\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.656412 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a42c7819-a52b-4dbc-bbe5-528c00fb6d95-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.804006 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.959809 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s4hj\" (UniqueName: \"kubernetes.io/projected/7df8bce2-d9fd-4875-8af0-86942949d93d-kube-api-access-5s4hj\") pod \"7df8bce2-d9fd-4875-8af0-86942949d93d\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.959849 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-catalog-content\") pod \"7df8bce2-d9fd-4875-8af0-86942949d93d\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.959897 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-utilities\") pod \"7df8bce2-d9fd-4875-8af0-86942949d93d\" (UID: \"7df8bce2-d9fd-4875-8af0-86942949d93d\") " Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.960904 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-utilities" (OuterVolumeSpecName: "utilities") pod "7df8bce2-d9fd-4875-8af0-86942949d93d" (UID: "7df8bce2-d9fd-4875-8af0-86942949d93d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:06 crc kubenswrapper[4696]: I1125 10:40:06.962555 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df8bce2-d9fd-4875-8af0-86942949d93d-kube-api-access-5s4hj" (OuterVolumeSpecName: "kube-api-access-5s4hj") pod "7df8bce2-d9fd-4875-8af0-86942949d93d" (UID: "7df8bce2-d9fd-4875-8af0-86942949d93d"). InnerVolumeSpecName "kube-api-access-5s4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.007308 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7df8bce2-d9fd-4875-8af0-86942949d93d" (UID: "7df8bce2-d9fd-4875-8af0-86942949d93d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.033905 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.061144 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s4hj\" (UniqueName: \"kubernetes.io/projected/7df8bce2-d9fd-4875-8af0-86942949d93d-kube-api-access-5s4hj\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.061197 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.061244 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7df8bce2-d9fd-4875-8af0-86942949d93d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.078448 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.078485 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88drx" event={"ID":"a42c7819-a52b-4dbc-bbe5-528c00fb6d95","Type":"ContainerDied","Data":"78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9"} Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.078514 4696 scope.go:117] "RemoveContainer" containerID="78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.078720 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-88drx" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.079824 4696 generic.go:334] "Generic (PLEG): container finished" podID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerID="78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9" exitCode=0 Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.079925 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-88drx" event={"ID":"a42c7819-a52b-4dbc-bbe5-528c00fb6d95","Type":"ContainerDied","Data":"8f35ab4fc84000675215ca50c5526171b08a605550b7737889065b784194ab6f"} Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.083038 4696 generic.go:334] "Generic (PLEG): container finished" podID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerID="4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72" exitCode=0 Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.083097 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerDied","Data":"4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72"} Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.083149 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvmp4" event={"ID":"7df8bce2-d9fd-4875-8af0-86942949d93d","Type":"ContainerDied","Data":"62ae701cac36ff8537e4ff13c02158ae2a333fe137040ee3e8495cdc6b2bac4c"} Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.083392 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvmp4" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.099630 4696 scope.go:117] "RemoveContainer" containerID="83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.123169 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-88drx"] Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.126559 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-88drx"] Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.131897 4696 scope.go:117] "RemoveContainer" containerID="39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.132197 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvmp4"] Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.138542 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jvmp4"] Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.146891 4696 scope.go:117] "RemoveContainer" containerID="78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9" Nov 25 10:40:07 crc kubenswrapper[4696]: E1125 10:40:07.148157 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9\": container with ID starting with 78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9 not found: ID does not exist" containerID="78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.148198 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9"} err="failed to get container status \"78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9\": rpc error: code = NotFound desc = could not find container \"78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9\": container with ID starting with 78c8c1a452dc0b2df58e02f169ccdeba0d07c386bea4665304dd0ebb05ba84d9 not found: ID does not exist" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.148245 4696 scope.go:117] "RemoveContainer" containerID="83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1" Nov 25 10:40:07 crc kubenswrapper[4696]: E1125 10:40:07.148912 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1\": container with ID starting with 83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1 not found: ID does not exist" containerID="83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.148940 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1"} err="failed to get container status \"83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1\": rpc error: code = NotFound desc = could not find container \"83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1\": container with ID starting with 83b6a4e58649ec50aa3def7ef66a2b3cb6116e561c69b964417c9b5ff386f7d1 not found: ID does not exist" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.148960 4696 scope.go:117] "RemoveContainer" containerID="39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25" Nov 25 10:40:07 crc kubenswrapper[4696]: E1125 10:40:07.149511 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25\": container with ID starting with 39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25 not found: ID does not exist" containerID="39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.149576 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25"} err="failed to get container status \"39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25\": rpc error: code = NotFound desc = could not find container \"39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25\": container with ID starting with 39acd74e61ce3aadd4cefedc4ccdbf7399d6b5de936cf4c9031a600baade1c25 not found: ID does not exist" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.149596 4696 scope.go:117] "RemoveContainer" containerID="4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.161429 4696 scope.go:117] "RemoveContainer" containerID="0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.180193 4696 scope.go:117] "RemoveContainer" containerID="834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.214532 4696 scope.go:117] "RemoveContainer" containerID="4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72" Nov 25 10:40:07 crc kubenswrapper[4696]: E1125 10:40:07.215141 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72\": container with ID starting with 4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72 not found: ID does not exist" containerID="4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.215264 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72"} err="failed to get container status \"4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72\": rpc error: code = NotFound desc = could not find container \"4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72\": container with ID starting with 4c075bb489c86f038c4523410d47ba25dfa1cc17e114268836c1113e5a965d72 not found: ID does not exist" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.215363 4696 scope.go:117] "RemoveContainer" containerID="0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f" Nov 25 10:40:07 crc kubenswrapper[4696]: E1125 10:40:07.215740 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f\": container with ID starting with 0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f not found: ID does not exist" containerID="0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.215841 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f"} err="failed to get container status \"0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f\": rpc error: code = NotFound desc = could not find container \"0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f\": container with ID starting with 0ea06db8324e56f392e1bc92346a9ca3527544517624236a008c9378c5c4521f not found: ID does not exist" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.215949 4696 scope.go:117] "RemoveContainer" containerID="834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd" Nov 25 10:40:07 crc kubenswrapper[4696]: E1125 10:40:07.216244 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd\": container with ID starting with 834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd not found: ID does not exist" containerID="834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd" Nov 25 10:40:07 crc kubenswrapper[4696]: I1125 10:40:07.216363 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd"} err="failed to get container status \"834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd\": rpc error: code = NotFound desc = could not find container \"834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd\": container with ID starting with 834517f34f2651ed675b1b400f1aa2a9c5435c55726e62fb377ec20dd5297edd not found: ID does not exist" Nov 25 10:40:08 crc kubenswrapper[4696]: I1125 10:40:08.053885 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" path="/var/lib/kubelet/pods/7df8bce2-d9fd-4875-8af0-86942949d93d/volumes" Nov 25 10:40:08 crc kubenswrapper[4696]: I1125 10:40:08.054781 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" path="/var/lib/kubelet/pods/a42c7819-a52b-4dbc-bbe5-528c00fb6d95/volumes" Nov 25 10:40:08 crc kubenswrapper[4696]: I1125 10:40:08.865016 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8fmqn"] Nov 25 10:40:08 crc kubenswrapper[4696]: I1125 10:40:08.865363 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8fmqn" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="registry-server" containerID="cri-o://e5acb745c267495a4abed05e8e8b56ef598db11ca66814416f76085992390b8d" gracePeriod=2 Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.099203 4696 generic.go:334] "Generic (PLEG): container finished" podID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerID="e5acb745c267495a4abed05e8e8b56ef598db11ca66814416f76085992390b8d" exitCode=0 Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.099480 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8fmqn" event={"ID":"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01","Type":"ContainerDied","Data":"e5acb745c267495a4abed05e8e8b56ef598db11ca66814416f76085992390b8d"} Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.216115 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.287969 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-utilities\") pod \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.288024 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxdss\" (UniqueName: \"kubernetes.io/projected/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-kube-api-access-xxdss\") pod \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.288181 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-catalog-content\") pod \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\" (UID: \"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01\") " Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.288868 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-utilities" (OuterVolumeSpecName: "utilities") pod "9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" (UID: "9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.290857 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.294803 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-kube-api-access-xxdss" (OuterVolumeSpecName: "kube-api-access-xxdss") pod "9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" (UID: "9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01"). InnerVolumeSpecName "kube-api-access-xxdss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.306464 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" (UID: "9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.392119 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:09 crc kubenswrapper[4696]: I1125 10:40:09.392166 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxdss\" (UniqueName: \"kubernetes.io/projected/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01-kube-api-access-xxdss\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.105214 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8fmqn" event={"ID":"9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01","Type":"ContainerDied","Data":"4b96d5f4ad26900decf934da743d5c90c205fddfdde2439cb3caa689a275dac2"} Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.105260 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8fmqn" Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.105618 4696 scope.go:117] "RemoveContainer" containerID="e5acb745c267495a4abed05e8e8b56ef598db11ca66814416f76085992390b8d" Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.131463 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8fmqn"] Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.134823 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8fmqn"] Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.135054 4696 scope.go:117] "RemoveContainer" containerID="c35578be44aed651f8504eeca8c0fe142b087b9b6f2bf70be5e7244bd48ab8d7" Nov 25 10:40:10 crc kubenswrapper[4696]: I1125 10:40:10.149968 4696 scope.go:117] "RemoveContainer" containerID="56f381eea2d3fd1652645b7e196531e080c388a8e3cf019ed99597e233b781d5" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.263911 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j6j5q"] Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.264121 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j6j5q" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="registry-server" containerID="cri-o://523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6" gracePeriod=2 Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.594117 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.716532 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-catalog-content\") pod \"1b409cdd-df72-4365-983d-852916369a5a\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.716689 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55gj2\" (UniqueName: \"kubernetes.io/projected/1b409cdd-df72-4365-983d-852916369a5a-kube-api-access-55gj2\") pod \"1b409cdd-df72-4365-983d-852916369a5a\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.716720 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-utilities\") pod \"1b409cdd-df72-4365-983d-852916369a5a\" (UID: \"1b409cdd-df72-4365-983d-852916369a5a\") " Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.718068 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-utilities" (OuterVolumeSpecName: "utilities") pod "1b409cdd-df72-4365-983d-852916369a5a" (UID: "1b409cdd-df72-4365-983d-852916369a5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.721999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b409cdd-df72-4365-983d-852916369a5a-kube-api-access-55gj2" (OuterVolumeSpecName: "kube-api-access-55gj2") pod "1b409cdd-df72-4365-983d-852916369a5a" (UID: "1b409cdd-df72-4365-983d-852916369a5a"). InnerVolumeSpecName "kube-api-access-55gj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.813642 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b409cdd-df72-4365-983d-852916369a5a" (UID: "1b409cdd-df72-4365-983d-852916369a5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.819331 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55gj2\" (UniqueName: \"kubernetes.io/projected/1b409cdd-df72-4365-983d-852916369a5a-kube-api-access-55gj2\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.819631 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:11 crc kubenswrapper[4696]: I1125 10:40:11.819738 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b409cdd-df72-4365-983d-852916369a5a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.048444 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" path="/var/lib/kubelet/pods/9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01/volumes" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.117185 4696 generic.go:334] "Generic (PLEG): container finished" podID="1b409cdd-df72-4365-983d-852916369a5a" containerID="523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6" exitCode=0 Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.117227 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6j5q" event={"ID":"1b409cdd-df72-4365-983d-852916369a5a","Type":"ContainerDied","Data":"523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6"} Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.117278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6j5q" event={"ID":"1b409cdd-df72-4365-983d-852916369a5a","Type":"ContainerDied","Data":"6d392815d876c681756e1572fa45520540a5bfc331c5de60315ed094538d0a3d"} Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.117294 4696 scope.go:117] "RemoveContainer" containerID="523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.117477 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6j5q" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.136472 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j6j5q"] Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.137203 4696 scope.go:117] "RemoveContainer" containerID="0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.140063 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j6j5q"] Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.157510 4696 scope.go:117] "RemoveContainer" containerID="2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.176007 4696 scope.go:117] "RemoveContainer" containerID="523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6" Nov 25 10:40:12 crc kubenswrapper[4696]: E1125 10:40:12.176483 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6\": container with ID starting with 523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6 not found: ID does not exist" containerID="523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.176519 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6"} err="failed to get container status \"523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6\": rpc error: code = NotFound desc = could not find container \"523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6\": container with ID starting with 523ce8df2cd75db619d0d94500247ccce105a1bf3056a2d2e0309539ea94afe6 not found: ID does not exist" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.176546 4696 scope.go:117] "RemoveContainer" containerID="0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14" Nov 25 10:40:12 crc kubenswrapper[4696]: E1125 10:40:12.176979 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14\": container with ID starting with 0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14 not found: ID does not exist" containerID="0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.177007 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14"} err="failed to get container status \"0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14\": rpc error: code = NotFound desc = could not find container \"0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14\": container with ID starting with 0ea3577ec1a2aaa38a28c59a04cb3c560ccd7bb0ee441433b583412204369b14 not found: ID does not exist" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.177023 4696 scope.go:117] "RemoveContainer" containerID="2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d" Nov 25 10:40:12 crc kubenswrapper[4696]: E1125 10:40:12.177263 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d\": container with ID starting with 2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d not found: ID does not exist" containerID="2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d" Nov 25 10:40:12 crc kubenswrapper[4696]: I1125 10:40:12.177326 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d"} err="failed to get container status \"2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d\": rpc error: code = NotFound desc = could not find container \"2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d\": container with ID starting with 2da963b39196cea0fc4fe9243a196d0e83c10495683a8e96c85d43d9a530306d not found: ID does not exist" Nov 25 10:40:14 crc kubenswrapper[4696]: I1125 10:40:14.051787 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b409cdd-df72-4365-983d-852916369a5a" path="/var/lib/kubelet/pods/1b409cdd-df72-4365-983d-852916369a5a/volumes" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.143970 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerName="oauth-openshift" containerID="cri-o://781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c" gracePeriod=15 Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.506527 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547354 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5cd7676b57-6nncv"] Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547599 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547619 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547630 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55e2607-be7d-4211-9b08-cf6905029542" containerName="pruner" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547638 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55e2607-be7d-4211-9b08-cf6905029542" containerName="pruner" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547650 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547657 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547687 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547695 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547706 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bb6022e-d54f-4374-bc8b-4dd422cd0a36" containerName="pruner" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547714 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bb6022e-d54f-4374-bc8b-4dd422cd0a36" containerName="pruner" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547726 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6edd3c-d469-4a55-a9b8-2a3087248db6" containerName="collect-profiles" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547733 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6edd3c-d469-4a55-a9b8-2a3087248db6" containerName="collect-profiles" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547742 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547749 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547760 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547767 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547776 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547783 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547794 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547803 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547815 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547822 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547835 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerName="oauth-openshift" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547843 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerName="oauth-openshift" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547852 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547859 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="extract-utilities" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547871 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547879 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547894 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547901 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="extract-content" Nov 25 10:40:29 crc kubenswrapper[4696]: E1125 10:40:29.547915 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.547922 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548034 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b409cdd-df72-4365-983d-852916369a5a" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548049 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9077ef6e-ed86-4e0c-a926-fe8b9d1b9a01" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548058 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42c7819-a52b-4dbc-bbe5-528c00fb6d95" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548067 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6edd3c-d469-4a55-a9b8-2a3087248db6" containerName="collect-profiles" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548074 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerName="oauth-openshift" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548084 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df8bce2-d9fd-4875-8af0-86942949d93d" containerName="registry-server" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548099 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e55e2607-be7d-4211-9b08-cf6905029542" containerName="pruner" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548108 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bb6022e-d54f-4374-bc8b-4dd422cd0a36" containerName="pruner" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.548580 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.563754 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5cd7676b57-6nncv"] Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.639893 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-dir\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.639960 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-provider-selection\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.639992 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-serving-cert\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640022 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-trusted-ca-bundle\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640047 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8jsw\" (UniqueName: \"kubernetes.io/projected/2e1aecfe-07a2-4772-bab2-52b5f5fed397-kube-api-access-q8jsw\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640081 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-idp-0-file-data\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640100 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-service-ca\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640121 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-login\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640154 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-router-certs\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640181 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-error\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640210 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-session\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640234 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-policies\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640273 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-cliconfig\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640296 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-ocp-branding-template\") pod \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\" (UID: \"2e1aecfe-07a2-4772-bab2-52b5f5fed397\") " Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.640770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.641128 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.641142 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.641424 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.641981 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.647317 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.648011 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.648220 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.649017 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.648912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.649516 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.650886 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e1aecfe-07a2-4772-bab2-52b5f5fed397-kube-api-access-q8jsw" (OuterVolumeSpecName: "kube-api-access-q8jsw") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "kube-api-access-q8jsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.651037 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.652146 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2e1aecfe-07a2-4772-bab2-52b5f5fed397" (UID: "2e1aecfe-07a2-4772-bab2-52b5f5fed397"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.741935 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.741987 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742008 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-login\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742035 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-audit-policies\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742066 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-session\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742097 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-error\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742121 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742144 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8jqx\" (UniqueName: \"kubernetes.io/projected/38fa1d59-f4aa-4f1c-a072-01225db2e541-kube-api-access-z8jqx\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742170 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742236 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742288 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742352 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742381 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/38fa1d59-f4aa-4f1c-a072-01225db2e541-audit-dir\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742410 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742487 4696 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742500 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742511 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742521 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742531 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8jsw\" (UniqueName: \"kubernetes.io/projected/2e1aecfe-07a2-4772-bab2-52b5f5fed397-kube-api-access-q8jsw\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742541 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742550 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742559 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742567 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742577 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742586 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742594 4696 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742602 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.742610 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e1aecfe-07a2-4772-bab2-52b5f5fed397-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843492 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843575 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8jqx\" (UniqueName: \"kubernetes.io/projected/38fa1d59-f4aa-4f1c-a072-01225db2e541-kube-api-access-z8jqx\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843623 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843715 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843756 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843818 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843859 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/38fa1d59-f4aa-4f1c-a072-01225db2e541-audit-dir\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843897 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843935 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.843982 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844021 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-login\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844064 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-audit-policies\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844104 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-session\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-error\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844395 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844520 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/38fa1d59-f4aa-4f1c-a072-01225db2e541-audit-dir\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.844781 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.845261 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-audit-policies\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.845649 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.847532 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.847600 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-login\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.847922 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.848371 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.849112 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.849603 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.849992 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-system-session\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.850544 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/38fa1d59-f4aa-4f1c-a072-01225db2e541-v4-0-config-user-template-error\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:29 crc kubenswrapper[4696]: I1125 10:40:29.863188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8jqx\" (UniqueName: \"kubernetes.io/projected/38fa1d59-f4aa-4f1c-a072-01225db2e541-kube-api-access-z8jqx\") pod \"oauth-openshift-5cd7676b57-6nncv\" (UID: \"38fa1d59-f4aa-4f1c-a072-01225db2e541\") " pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.162992 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.222420 4696 generic.go:334] "Generic (PLEG): container finished" podID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" containerID="781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c" exitCode=0 Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.222490 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" event={"ID":"2e1aecfe-07a2-4772-bab2-52b5f5fed397","Type":"ContainerDied","Data":"781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c"} Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.222527 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.222568 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nvzm5" event={"ID":"2e1aecfe-07a2-4772-bab2-52b5f5fed397","Type":"ContainerDied","Data":"8fa15718bc52862248ccf9428cd3fa22641d1f3c619fbf39b18188f9b744da76"} Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.222596 4696 scope.go:117] "RemoveContainer" containerID="781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.247534 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nvzm5"] Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.255262 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nvzm5"] Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.296181 4696 scope.go:117] "RemoveContainer" containerID="781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c" Nov 25 10:40:30 crc kubenswrapper[4696]: E1125 10:40:30.296592 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c\": container with ID starting with 781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c not found: ID does not exist" containerID="781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.296620 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c"} err="failed to get container status \"781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c\": rpc error: code = NotFound desc = could not find container \"781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c\": container with ID starting with 781c909b6a5186f35d414221b1f18af485cf294ba8f5a72fb3c8962fb5d1f35c not found: ID does not exist" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.630983 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5cd7676b57-6nncv"] Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.801549 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.801598 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.801643 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.802236 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:40:30 crc kubenswrapper[4696]: I1125 10:40:30.802299 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051" gracePeriod=600 Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.228635 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051" exitCode=0 Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.228919 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051"} Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.228944 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"1fb36885e86a8194f5415b7d52233333dec593850e5cf8d383f363bb7a6ad19c"} Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.231643 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" event={"ID":"38fa1d59-f4aa-4f1c-a072-01225db2e541","Type":"ContainerStarted","Data":"139f2a31ef545ac00c918aae53957d5a3d8754df0e02092680a33c13318e9ca3"} Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.231678 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" event={"ID":"38fa1d59-f4aa-4f1c-a072-01225db2e541","Type":"ContainerStarted","Data":"e93b10bf8048995fc4ed553546b03edeaa9afabc3430a2e41246787401631323"} Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.232256 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:31 crc kubenswrapper[4696]: I1125 10:40:31.241313 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" Nov 25 10:40:32 crc kubenswrapper[4696]: I1125 10:40:32.049827 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e1aecfe-07a2-4772-bab2-52b5f5fed397" path="/var/lib/kubelet/pods/2e1aecfe-07a2-4772-bab2-52b5f5fed397/volumes" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.110350 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5cd7676b57-6nncv" podStartSLOduration=56.110332961 podStartE2EDuration="56.110332961s" podCreationTimestamp="2025-11-25 10:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:40:31.267564562 +0000 UTC m=+228.081181159" watchObservedRunningTime="2025-11-25 10:41:00.110332961 +0000 UTC m=+256.923949538" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.113583 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nd9j4"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.113836 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nd9j4" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="registry-server" containerID="cri-o://52e56e9d94d3d4d5a333fcf1797220f56c4562b119f5a797de0b6df0d1732c24" gracePeriod=30 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.130222 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tpmx6"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.130605 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tpmx6" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="registry-server" containerID="cri-o://88e90f3f438b54c08800951e7cd9fe24176256260937938c2e2663352983cbcb" gracePeriod=30 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.134887 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkn7w"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.135107 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" containerID="cri-o://26b98ed90897186f3e1061344ddb4bbbf62c8304a33fd262ac83b0d6675c6855" gracePeriod=30 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.154141 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kvxk"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.154722 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8kvxk" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="registry-server" containerID="cri-o://f3d5ec472d0f2280f4323178a595cebde0cd524ff043f0477b66801ffd4b8c0d" gracePeriod=30 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.165494 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52nv8"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.166071 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-52nv8" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="registry-server" containerID="cri-o://4bb5219ac9ca64c7654c2102a570cc66aa7d2df7e73eee7431d2fdf2d8ebc5f9" gracePeriod=30 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.188847 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrn8c"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.190147 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.211655 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqgtv\" (UniqueName: \"kubernetes.io/projected/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-kube-api-access-fqgtv\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.211749 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.211791 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.216478 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrn8c"] Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.312511 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.312588 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqgtv\" (UniqueName: \"kubernetes.io/projected/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-kube-api-access-fqgtv\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.312628 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.314639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.319543 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.343222 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqgtv\" (UniqueName: \"kubernetes.io/projected/e77b3a6e-c1e9-44ef-806f-6adcba39fe31-kube-api-access-fqgtv\") pod \"marketplace-operator-79b997595-xrn8c\" (UID: \"e77b3a6e-c1e9-44ef-806f-6adcba39fe31\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.417041 4696 generic.go:334] "Generic (PLEG): container finished" podID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerID="f3d5ec472d0f2280f4323178a595cebde0cd524ff043f0477b66801ffd4b8c0d" exitCode=0 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.417120 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kvxk" event={"ID":"1150bbc8-8968-47f9-be96-81e1de5b8043","Type":"ContainerDied","Data":"f3d5ec472d0f2280f4323178a595cebde0cd524ff043f0477b66801ffd4b8c0d"} Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.431450 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerID="52e56e9d94d3d4d5a333fcf1797220f56c4562b119f5a797de0b6df0d1732c24" exitCode=0 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.431539 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd9j4" event={"ID":"d6b24f5f-89c0-4078-be5d-8739b554e76e","Type":"ContainerDied","Data":"52e56e9d94d3d4d5a333fcf1797220f56c4562b119f5a797de0b6df0d1732c24"} Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.440339 4696 generic.go:334] "Generic (PLEG): container finished" podID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerID="4bb5219ac9ca64c7654c2102a570cc66aa7d2df7e73eee7431d2fdf2d8ebc5f9" exitCode=0 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.440400 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerDied","Data":"4bb5219ac9ca64c7654c2102a570cc66aa7d2df7e73eee7431d2fdf2d8ebc5f9"} Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.441874 4696 generic.go:334] "Generic (PLEG): container finished" podID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerID="26b98ed90897186f3e1061344ddb4bbbf62c8304a33fd262ac83b0d6675c6855" exitCode=0 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.441926 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" event={"ID":"78279129-b25e-4e7c-b563-6043c9ff4f88","Type":"ContainerDied","Data":"26b98ed90897186f3e1061344ddb4bbbf62c8304a33fd262ac83b0d6675c6855"} Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.443646 4696 generic.go:334] "Generic (PLEG): container finished" podID="46840e91-cc38-4a5c-beca-96f3735590c4" containerID="88e90f3f438b54c08800951e7cd9fe24176256260937938c2e2663352983cbcb" exitCode=0 Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.443691 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpmx6" event={"ID":"46840e91-cc38-4a5c-beca-96f3735590c4","Type":"ContainerDied","Data":"88e90f3f438b54c08800951e7cd9fe24176256260937938c2e2663352983cbcb"} Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.454271 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.592827 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.621840 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-catalog-content\") pod \"46840e91-cc38-4a5c-beca-96f3735590c4\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.621902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-utilities\") pod \"46840e91-cc38-4a5c-beca-96f3735590c4\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.621928 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrcx5\" (UniqueName: \"kubernetes.io/projected/46840e91-cc38-4a5c-beca-96f3735590c4-kube-api-access-qrcx5\") pod \"46840e91-cc38-4a5c-beca-96f3735590c4\" (UID: \"46840e91-cc38-4a5c-beca-96f3735590c4\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.624880 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-utilities" (OuterVolumeSpecName: "utilities") pod "46840e91-cc38-4a5c-beca-96f3735590c4" (UID: "46840e91-cc38-4a5c-beca-96f3735590c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.632069 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46840e91-cc38-4a5c-beca-96f3735590c4-kube-api-access-qrcx5" (OuterVolumeSpecName: "kube-api-access-qrcx5") pod "46840e91-cc38-4a5c-beca-96f3735590c4" (UID: "46840e91-cc38-4a5c-beca-96f3735590c4"). InnerVolumeSpecName "kube-api-access-qrcx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.678470 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.703211 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.703556 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46840e91-cc38-4a5c-beca-96f3735590c4" (UID: "46840e91-cc38-4a5c-beca-96f3735590c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.713398 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.730695 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.730756 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46840e91-cc38-4a5c-beca-96f3735590c4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.730780 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrcx5\" (UniqueName: \"kubernetes.io/projected/46840e91-cc38-4a5c-beca-96f3735590c4-kube-api-access-qrcx5\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.770007 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831803 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-utilities\") pod \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831858 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q99qr\" (UniqueName: \"kubernetes.io/projected/1150bbc8-8968-47f9-be96-81e1de5b8043-kube-api-access-q99qr\") pod \"1150bbc8-8968-47f9-be96-81e1de5b8043\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831890 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-catalog-content\") pod \"d6b24f5f-89c0-4078-be5d-8739b554e76e\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831927 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-catalog-content\") pod \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831945 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-utilities\") pod \"1150bbc8-8968-47f9-be96-81e1de5b8043\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831960 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5hnq\" (UniqueName: \"kubernetes.io/projected/d6b24f5f-89c0-4078-be5d-8739b554e76e-kube-api-access-z5hnq\") pod \"d6b24f5f-89c0-4078-be5d-8739b554e76e\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.831983 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k58vp\" (UniqueName: \"kubernetes.io/projected/2d0338a7-bada-47ba-a1b9-21fd05a46f55-kube-api-access-k58vp\") pod \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\" (UID: \"2d0338a7-bada-47ba-a1b9-21fd05a46f55\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.832002 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-catalog-content\") pod \"1150bbc8-8968-47f9-be96-81e1de5b8043\" (UID: \"1150bbc8-8968-47f9-be96-81e1de5b8043\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.832072 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-utilities\") pod \"d6b24f5f-89c0-4078-be5d-8739b554e76e\" (UID: \"d6b24f5f-89c0-4078-be5d-8739b554e76e\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.832089 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-operator-metrics\") pod \"78279129-b25e-4e7c-b563-6043c9ff4f88\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.833741 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-utilities" (OuterVolumeSpecName: "utilities") pod "d6b24f5f-89c0-4078-be5d-8739b554e76e" (UID: "d6b24f5f-89c0-4078-be5d-8739b554e76e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.834147 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-utilities" (OuterVolumeSpecName: "utilities") pod "1150bbc8-8968-47f9-be96-81e1de5b8043" (UID: "1150bbc8-8968-47f9-be96-81e1de5b8043"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.834319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-utilities" (OuterVolumeSpecName: "utilities") pod "2d0338a7-bada-47ba-a1b9-21fd05a46f55" (UID: "2d0338a7-bada-47ba-a1b9-21fd05a46f55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.841320 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1150bbc8-8968-47f9-be96-81e1de5b8043-kube-api-access-q99qr" (OuterVolumeSpecName: "kube-api-access-q99qr") pod "1150bbc8-8968-47f9-be96-81e1de5b8043" (UID: "1150bbc8-8968-47f9-be96-81e1de5b8043"). InnerVolumeSpecName "kube-api-access-q99qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.841447 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d0338a7-bada-47ba-a1b9-21fd05a46f55-kube-api-access-k58vp" (OuterVolumeSpecName: "kube-api-access-k58vp") pod "2d0338a7-bada-47ba-a1b9-21fd05a46f55" (UID: "2d0338a7-bada-47ba-a1b9-21fd05a46f55"). InnerVolumeSpecName "kube-api-access-k58vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.842576 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b24f5f-89c0-4078-be5d-8739b554e76e-kube-api-access-z5hnq" (OuterVolumeSpecName: "kube-api-access-z5hnq") pod "d6b24f5f-89c0-4078-be5d-8739b554e76e" (UID: "d6b24f5f-89c0-4078-be5d-8739b554e76e"). InnerVolumeSpecName "kube-api-access-z5hnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.844582 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "78279129-b25e-4e7c-b563-6043c9ff4f88" (UID: "78279129-b25e-4e7c-b563-6043c9ff4f88"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.854772 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1150bbc8-8968-47f9-be96-81e1de5b8043" (UID: "1150bbc8-8968-47f9-be96-81e1de5b8043"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.887441 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6b24f5f-89c0-4078-be5d-8739b554e76e" (UID: "d6b24f5f-89c0-4078-be5d-8739b554e76e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.929105 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d0338a7-bada-47ba-a1b9-21fd05a46f55" (UID: "2d0338a7-bada-47ba-a1b9-21fd05a46f55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933034 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qhfz\" (UniqueName: \"kubernetes.io/projected/78279129-b25e-4e7c-b563-6043c9ff4f88-kube-api-access-2qhfz\") pod \"78279129-b25e-4e7c-b563-6043c9ff4f88\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933193 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-trusted-ca\") pod \"78279129-b25e-4e7c-b563-6043c9ff4f88\" (UID: \"78279129-b25e-4e7c-b563-6043c9ff4f88\") " Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933361 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k58vp\" (UniqueName: \"kubernetes.io/projected/2d0338a7-bada-47ba-a1b9-21fd05a46f55-kube-api-access-k58vp\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933379 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933392 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933406 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933418 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933429 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q99qr\" (UniqueName: \"kubernetes.io/projected/1150bbc8-8968-47f9-be96-81e1de5b8043-kube-api-access-q99qr\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933440 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b24f5f-89c0-4078-be5d-8739b554e76e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933452 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d0338a7-bada-47ba-a1b9-21fd05a46f55-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933463 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5hnq\" (UniqueName: \"kubernetes.io/projected/d6b24f5f-89c0-4078-be5d-8739b554e76e-kube-api-access-z5hnq\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933474 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1150bbc8-8968-47f9-be96-81e1de5b8043-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.933930 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "78279129-b25e-4e7c-b563-6043c9ff4f88" (UID: "78279129-b25e-4e7c-b563-6043c9ff4f88"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.936321 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78279129-b25e-4e7c-b563-6043c9ff4f88-kube-api-access-2qhfz" (OuterVolumeSpecName: "kube-api-access-2qhfz") pod "78279129-b25e-4e7c-b563-6043c9ff4f88" (UID: "78279129-b25e-4e7c-b563-6043c9ff4f88"). InnerVolumeSpecName "kube-api-access-2qhfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:41:00 crc kubenswrapper[4696]: I1125 10:41:00.996676 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrn8c"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.034717 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78279129-b25e-4e7c-b563-6043c9ff4f88-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.034755 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qhfz\" (UniqueName: \"kubernetes.io/projected/78279129-b25e-4e7c-b563-6043c9ff4f88-kube-api-access-2qhfz\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.451241 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpmx6" event={"ID":"46840e91-cc38-4a5c-beca-96f3735590c4","Type":"ContainerDied","Data":"891ecc107ecb77c4c750eba7266c0cc140437b8dd700b0fb74d497919d5dd82f"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.451308 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpmx6" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.451323 4696 scope.go:117] "RemoveContainer" containerID="88e90f3f438b54c08800951e7cd9fe24176256260937938c2e2663352983cbcb" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.453396 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" event={"ID":"e77b3a6e-c1e9-44ef-806f-6adcba39fe31","Type":"ContainerStarted","Data":"38b63ce85fbb6cbc7fc753cb0cebb9ab50b7c027ea18b36395d9bba487f5a695"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.453427 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" event={"ID":"e77b3a6e-c1e9-44ef-806f-6adcba39fe31","Type":"ContainerStarted","Data":"fef3e2ef3b27ec0720ed3709e5263c685b1f36fa057fffe6bc3a38d65ee77dc8"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.453693 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.456277 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kvxk" event={"ID":"1150bbc8-8968-47f9-be96-81e1de5b8043","Type":"ContainerDied","Data":"1a600f4d43912673573a95cfd509d9842c2ff8fcfd26c117802a8003fac8e6cc"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.456511 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kvxk" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.459275 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.461321 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd9j4" event={"ID":"d6b24f5f-89c0-4078-be5d-8739b554e76e","Type":"ContainerDied","Data":"a916926c82d28802b4e4284e74165c4a61522692744fe42c7ce65c8d315005b2"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.461416 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd9j4" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.468004 4696 scope.go:117] "RemoveContainer" containerID="76ed17a3f8550e38520fc69ccb4edaa54e050e2c41d0db6ef023e163a870a423" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.470064 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52nv8" event={"ID":"2d0338a7-bada-47ba-a1b9-21fd05a46f55","Type":"ContainerDied","Data":"193af7cd19f9e48736658baff6c263402e220802dfcf6c03cc8fae9e526e8505"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.470243 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52nv8" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.475771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" event={"ID":"78279129-b25e-4e7c-b563-6043c9ff4f88","Type":"ContainerDied","Data":"e10f3008b1d2381f305ebce42392850e08eeb28679e31bc8c71d01a47cac3d67"} Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.475885 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkn7w" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.504384 4696 scope.go:117] "RemoveContainer" containerID="ddb0e1db24961ed8b91e5e64177cf72255e6668b196a9eb15f35edfd24019be6" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.522319 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" podStartSLOduration=1.522300228 podStartE2EDuration="1.522300228s" podCreationTimestamp="2025-11-25 10:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:41:01.491690825 +0000 UTC m=+258.305307412" watchObservedRunningTime="2025-11-25 10:41:01.522300228 +0000 UTC m=+258.335916805" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.524882 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tpmx6"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.531893 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tpmx6"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.547503 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nd9j4"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.550481 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nd9j4"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.557356 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkn7w"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.564168 4696 scope.go:117] "RemoveContainer" containerID="f3d5ec472d0f2280f4323178a595cebde0cd524ff043f0477b66801ffd4b8c0d" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.567504 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkn7w"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.575855 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kvxk"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.578943 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kvxk"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.587727 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52nv8"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.588690 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-52nv8"] Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.593498 4696 scope.go:117] "RemoveContainer" containerID="5f924c3562f5458f055a6456f870ac20e06d5a928d0a22f1d497411c08e0ce26" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.612092 4696 scope.go:117] "RemoveContainer" containerID="06e7ab4b39b6fbe4622bc213bcca677b6a6e6938ea5236eb7c0e80f647e9e1a8" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.625190 4696 scope.go:117] "RemoveContainer" containerID="52e56e9d94d3d4d5a333fcf1797220f56c4562b119f5a797de0b6df0d1732c24" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.638308 4696 scope.go:117] "RemoveContainer" containerID="47e5459d1a409d36eca96df2186bada0a84143b5724a0f67820b244f81bd6429" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.649895 4696 scope.go:117] "RemoveContainer" containerID="7fff65983db5a1d5a438cd3f871f6aab6e1ee0f1fc63eb4495e4cd55ebd19925" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.661527 4696 scope.go:117] "RemoveContainer" containerID="4bb5219ac9ca64c7654c2102a570cc66aa7d2df7e73eee7431d2fdf2d8ebc5f9" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.672889 4696 scope.go:117] "RemoveContainer" containerID="f4e30fe4862c4d4e2cc769b34ae0138d2513a4203e0f8e80afe8fb031de04918" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.685970 4696 scope.go:117] "RemoveContainer" containerID="b291ae23dab9bd3d9239d7b9bd9b2ffd2b170e08b8d7303ae58ba572f9a19c3d" Nov 25 10:41:01 crc kubenswrapper[4696]: I1125 10:41:01.700320 4696 scope.go:117] "RemoveContainer" containerID="26b98ed90897186f3e1061344ddb4bbbf62c8304a33fd262ac83b0d6675c6855" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.047469 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" path="/var/lib/kubelet/pods/1150bbc8-8968-47f9-be96-81e1de5b8043/volumes" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.048704 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" path="/var/lib/kubelet/pods/2d0338a7-bada-47ba-a1b9-21fd05a46f55/volumes" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.049398 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" path="/var/lib/kubelet/pods/46840e91-cc38-4a5c-beca-96f3735590c4/volumes" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.050636 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" path="/var/lib/kubelet/pods/78279129-b25e-4e7c-b563-6043c9ff4f88/volumes" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.051206 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" path="/var/lib/kubelet/pods/d6b24f5f-89c0-4078-be5d-8739b554e76e/volumes" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330357 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h9ssz"] Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330640 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330657 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330693 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330702 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330716 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330725 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330742 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330751 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330763 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330771 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330783 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330793 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330802 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330810 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330821 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330830 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330843 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330851 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330861 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330868 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="extract-utilities" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330878 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330887 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330900 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330909 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="extract-content" Nov 25 10:41:02 crc kubenswrapper[4696]: E1125 10:41:02.330920 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.330930 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.331050 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="46840e91-cc38-4a5c-beca-96f3735590c4" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.331072 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d0338a7-bada-47ba-a1b9-21fd05a46f55" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.331083 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b24f5f-89c0-4078-be5d-8739b554e76e" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.331093 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="78279129-b25e-4e7c-b563-6043c9ff4f88" containerName="marketplace-operator" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.331103 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1150bbc8-8968-47f9-be96-81e1de5b8043" containerName="registry-server" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.332183 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.337183 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.340950 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9ssz"] Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.350881 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df2ddc89-6af7-4142-a3bd-58856bf7eb77-utilities\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.350928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df2ddc89-6af7-4142-a3bd-58856bf7eb77-catalog-content\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.350956 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg8xg\" (UniqueName: \"kubernetes.io/projected/df2ddc89-6af7-4142-a3bd-58856bf7eb77-kube-api-access-sg8xg\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.451875 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df2ddc89-6af7-4142-a3bd-58856bf7eb77-utilities\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.451923 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df2ddc89-6af7-4142-a3bd-58856bf7eb77-catalog-content\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.451942 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg8xg\" (UniqueName: \"kubernetes.io/projected/df2ddc89-6af7-4142-a3bd-58856bf7eb77-kube-api-access-sg8xg\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.452521 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df2ddc89-6af7-4142-a3bd-58856bf7eb77-catalog-content\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.452538 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df2ddc89-6af7-4142-a3bd-58856bf7eb77-utilities\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.468255 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg8xg\" (UniqueName: \"kubernetes.io/projected/df2ddc89-6af7-4142-a3bd-58856bf7eb77-kube-api-access-sg8xg\") pod \"redhat-marketplace-h9ssz\" (UID: \"df2ddc89-6af7-4142-a3bd-58856bf7eb77\") " pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.540740 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7pxd4"] Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.542194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.544167 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.550393 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7pxd4"] Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.553092 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-utilities\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.553218 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb27q\" (UniqueName: \"kubernetes.io/projected/250727c1-0261-49a2-b93f-86ac9235b443-kube-api-access-rb27q\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.553340 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-catalog-content\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.654284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-catalog-content\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.654373 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-utilities\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.654415 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb27q\" (UniqueName: \"kubernetes.io/projected/250727c1-0261-49a2-b93f-86ac9235b443-kube-api-access-rb27q\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.655033 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.655226 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-catalog-content\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.655253 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-utilities\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.672550 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb27q\" (UniqueName: \"kubernetes.io/projected/250727c1-0261-49a2-b93f-86ac9235b443-kube-api-access-rb27q\") pod \"redhat-operators-7pxd4\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:02 crc kubenswrapper[4696]: I1125 10:41:02.866065 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.030611 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9ssz"] Nov 25 10:41:03 crc kubenswrapper[4696]: W1125 10:41:03.036823 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf2ddc89_6af7_4142_a3bd_58856bf7eb77.slice/crio-219a2442cac5dbbbe6d5983285410ec958bcecbc0e102350c8a71d7b5fbfa0d3 WatchSource:0}: Error finding container 219a2442cac5dbbbe6d5983285410ec958bcecbc0e102350c8a71d7b5fbfa0d3: Status 404 returned error can't find the container with id 219a2442cac5dbbbe6d5983285410ec958bcecbc0e102350c8a71d7b5fbfa0d3 Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.243174 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7pxd4"] Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.494976 4696 generic.go:334] "Generic (PLEG): container finished" podID="df2ddc89-6af7-4142-a3bd-58856bf7eb77" containerID="8560cdace32d868126682a080e8df6dd932dc295b41aef731d522cdc3eb05870" exitCode=0 Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.495091 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9ssz" event={"ID":"df2ddc89-6af7-4142-a3bd-58856bf7eb77","Type":"ContainerDied","Data":"8560cdace32d868126682a080e8df6dd932dc295b41aef731d522cdc3eb05870"} Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.495312 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9ssz" event={"ID":"df2ddc89-6af7-4142-a3bd-58856bf7eb77","Type":"ContainerStarted","Data":"219a2442cac5dbbbe6d5983285410ec958bcecbc0e102350c8a71d7b5fbfa0d3"} Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.498771 4696 generic.go:334] "Generic (PLEG): container finished" podID="250727c1-0261-49a2-b93f-86ac9235b443" containerID="0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2" exitCode=0 Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.498971 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerDied","Data":"0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2"} Nov 25 10:41:03 crc kubenswrapper[4696]: I1125 10:41:03.499011 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerStarted","Data":"8719bf798fbd5402f3a447f533da3d7cc029ae103ab9efca534d27904fd6dc48"} Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.505342 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerStarted","Data":"5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d"} Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.729161 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s9cff"] Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.730409 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.733798 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.740387 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s9cff"] Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.782404 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-utilities\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.782465 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-catalog-content\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.782501 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x696l\" (UniqueName: \"kubernetes.io/projected/de81aa8f-a3f2-42c0-8a6d-1756748ec575-kube-api-access-x696l\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.883923 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-utilities\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.883996 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-catalog-content\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.884065 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x696l\" (UniqueName: \"kubernetes.io/projected/de81aa8f-a3f2-42c0-8a6d-1756748ec575-kube-api-access-x696l\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.884428 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-utilities\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.884432 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-catalog-content\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.905455 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x696l\" (UniqueName: \"kubernetes.io/projected/de81aa8f-a3f2-42c0-8a6d-1756748ec575-kube-api-access-x696l\") pod \"certified-operators-s9cff\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.944941 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nbc7f"] Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.946279 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.948807 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.958810 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbc7f"] Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.985897 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrd5p\" (UniqueName: \"kubernetes.io/projected/4ba6a060-a9da-44c9-b1c8-450ea6075e18-kube-api-access-lrd5p\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.986038 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba6a060-a9da-44c9-b1c8-450ea6075e18-catalog-content\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:04 crc kubenswrapper[4696]: I1125 10:41:04.986270 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba6a060-a9da-44c9-b1c8-450ea6075e18-utilities\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.061927 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.087889 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba6a060-a9da-44c9-b1c8-450ea6075e18-catalog-content\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.087978 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba6a060-a9da-44c9-b1c8-450ea6075e18-utilities\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.088011 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrd5p\" (UniqueName: \"kubernetes.io/projected/4ba6a060-a9da-44c9-b1c8-450ea6075e18-kube-api-access-lrd5p\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.089725 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba6a060-a9da-44c9-b1c8-450ea6075e18-catalog-content\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.092191 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba6a060-a9da-44c9-b1c8-450ea6075e18-utilities\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.108877 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrd5p\" (UniqueName: \"kubernetes.io/projected/4ba6a060-a9da-44c9-b1c8-450ea6075e18-kube-api-access-lrd5p\") pod \"community-operators-nbc7f\" (UID: \"4ba6a060-a9da-44c9-b1c8-450ea6075e18\") " pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.268527 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.459543 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s9cff"] Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.513888 4696 generic.go:334] "Generic (PLEG): container finished" podID="df2ddc89-6af7-4142-a3bd-58856bf7eb77" containerID="90fb0bc7718b86f6d194a38f448a1bb383a7e35e0c20645635c171b91f3c4579" exitCode=0 Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.514053 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9ssz" event={"ID":"df2ddc89-6af7-4142-a3bd-58856bf7eb77","Type":"ContainerDied","Data":"90fb0bc7718b86f6d194a38f448a1bb383a7e35e0c20645635c171b91f3c4579"} Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.516395 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9cff" event={"ID":"de81aa8f-a3f2-42c0-8a6d-1756748ec575","Type":"ContainerStarted","Data":"e9a75391b6383a7a9a21c9771d5dc2e8cdbcd40ccd4cb4d9f3c1c0aac0eccf1e"} Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.521240 4696 generic.go:334] "Generic (PLEG): container finished" podID="250727c1-0261-49a2-b93f-86ac9235b443" containerID="5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d" exitCode=0 Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.521361 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerDied","Data":"5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d"} Nov 25 10:41:05 crc kubenswrapper[4696]: I1125 10:41:05.645518 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbc7f"] Nov 25 10:41:05 crc kubenswrapper[4696]: W1125 10:41:05.672128 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ba6a060_a9da_44c9_b1c8_450ea6075e18.slice/crio-0bd90c641a8605d7f581ec9281ac8dc7db6d1ba1d941139b518eb386e31a48c1 WatchSource:0}: Error finding container 0bd90c641a8605d7f581ec9281ac8dc7db6d1ba1d941139b518eb386e31a48c1: Status 404 returned error can't find the container with id 0bd90c641a8605d7f581ec9281ac8dc7db6d1ba1d941139b518eb386e31a48c1 Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.528034 4696 generic.go:334] "Generic (PLEG): container finished" podID="4ba6a060-a9da-44c9-b1c8-450ea6075e18" containerID="9d8643c5fdfb9c3f906b8e86f4db3b0a5cd5e4e17fbb581da14f8e935d7a329c" exitCode=0 Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.528079 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbc7f" event={"ID":"4ba6a060-a9da-44c9-b1c8-450ea6075e18","Type":"ContainerDied","Data":"9d8643c5fdfb9c3f906b8e86f4db3b0a5cd5e4e17fbb581da14f8e935d7a329c"} Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.530111 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbc7f" event={"ID":"4ba6a060-a9da-44c9-b1c8-450ea6075e18","Type":"ContainerStarted","Data":"0bd90c641a8605d7f581ec9281ac8dc7db6d1ba1d941139b518eb386e31a48c1"} Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.533985 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9ssz" event={"ID":"df2ddc89-6af7-4142-a3bd-58856bf7eb77","Type":"ContainerStarted","Data":"30bbe1330c4a58cb39fd843c2e2d3c004f865f8faed026181f93dbc0b1921c4b"} Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.535483 4696 generic.go:334] "Generic (PLEG): container finished" podID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerID="7d3d0db5ee1fcae0a8c538889d5767c8ec3f5bfaf96334825adfb498eb2036b6" exitCode=0 Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.535553 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9cff" event={"ID":"de81aa8f-a3f2-42c0-8a6d-1756748ec575","Type":"ContainerDied","Data":"7d3d0db5ee1fcae0a8c538889d5767c8ec3f5bfaf96334825adfb498eb2036b6"} Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.537950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerStarted","Data":"df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd"} Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.593356 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h9ssz" podStartSLOduration=2.038839447 podStartE2EDuration="4.593335827s" podCreationTimestamp="2025-11-25 10:41:02 +0000 UTC" firstStartedPulling="2025-11-25 10:41:03.496800748 +0000 UTC m=+260.310417335" lastFinishedPulling="2025-11-25 10:41:06.051297128 +0000 UTC m=+262.864913715" observedRunningTime="2025-11-25 10:41:06.592770559 +0000 UTC m=+263.406387156" watchObservedRunningTime="2025-11-25 10:41:06.593335827 +0000 UTC m=+263.406952414" Nov 25 10:41:06 crc kubenswrapper[4696]: I1125 10:41:06.615003 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7pxd4" podStartSLOduration=2.036670114 podStartE2EDuration="4.6149832s" podCreationTimestamp="2025-11-25 10:41:02 +0000 UTC" firstStartedPulling="2025-11-25 10:41:03.501059268 +0000 UTC m=+260.314675855" lastFinishedPulling="2025-11-25 10:41:06.079372354 +0000 UTC m=+262.892988941" observedRunningTime="2025-11-25 10:41:06.613945257 +0000 UTC m=+263.427561844" watchObservedRunningTime="2025-11-25 10:41:06.6149832 +0000 UTC m=+263.428599787" Nov 25 10:41:07 crc kubenswrapper[4696]: I1125 10:41:07.544059 4696 generic.go:334] "Generic (PLEG): container finished" podID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerID="8f5891f5a89280ba498edb75fe8c6a5f8af52c809a58bf341afbb30ab167864e" exitCode=0 Nov 25 10:41:07 crc kubenswrapper[4696]: I1125 10:41:07.545383 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9cff" event={"ID":"de81aa8f-a3f2-42c0-8a6d-1756748ec575","Type":"ContainerDied","Data":"8f5891f5a89280ba498edb75fe8c6a5f8af52c809a58bf341afbb30ab167864e"} Nov 25 10:41:07 crc kubenswrapper[4696]: I1125 10:41:07.551048 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbc7f" event={"ID":"4ba6a060-a9da-44c9-b1c8-450ea6075e18","Type":"ContainerStarted","Data":"45ded48383d5c57f9e753101dc34c09e935be1557a2a3e478ba109c6393c9b08"} Nov 25 10:41:08 crc kubenswrapper[4696]: I1125 10:41:08.557774 4696 generic.go:334] "Generic (PLEG): container finished" podID="4ba6a060-a9da-44c9-b1c8-450ea6075e18" containerID="45ded48383d5c57f9e753101dc34c09e935be1557a2a3e478ba109c6393c9b08" exitCode=0 Nov 25 10:41:08 crc kubenswrapper[4696]: I1125 10:41:08.557859 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbc7f" event={"ID":"4ba6a060-a9da-44c9-b1c8-450ea6075e18","Type":"ContainerDied","Data":"45ded48383d5c57f9e753101dc34c09e935be1557a2a3e478ba109c6393c9b08"} Nov 25 10:41:08 crc kubenswrapper[4696]: I1125 10:41:08.561947 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9cff" event={"ID":"de81aa8f-a3f2-42c0-8a6d-1756748ec575","Type":"ContainerStarted","Data":"b47bd19e20eead4570cbc57d0db67874828e25eb19630cea77d764efc212eb59"} Nov 25 10:41:08 crc kubenswrapper[4696]: I1125 10:41:08.592691 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s9cff" podStartSLOduration=2.861310489 podStartE2EDuration="4.592650709s" podCreationTimestamp="2025-11-25 10:41:04 +0000 UTC" firstStartedPulling="2025-11-25 10:41:06.536560953 +0000 UTC m=+263.350177540" lastFinishedPulling="2025-11-25 10:41:08.267901173 +0000 UTC m=+265.081517760" observedRunningTime="2025-11-25 10:41:08.590637855 +0000 UTC m=+265.404254442" watchObservedRunningTime="2025-11-25 10:41:08.592650709 +0000 UTC m=+265.406267296" Nov 25 10:41:10 crc kubenswrapper[4696]: I1125 10:41:10.574735 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbc7f" event={"ID":"4ba6a060-a9da-44c9-b1c8-450ea6075e18","Type":"ContainerStarted","Data":"c6a2ba5bc8087d418267c1f5f730336d9d4a5ce360867f678e83b734277d8c03"} Nov 25 10:41:10 crc kubenswrapper[4696]: I1125 10:41:10.608006 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nbc7f" podStartSLOduration=4.143831415 podStartE2EDuration="6.607984667s" podCreationTimestamp="2025-11-25 10:41:04 +0000 UTC" firstStartedPulling="2025-11-25 10:41:06.529451239 +0000 UTC m=+263.343067826" lastFinishedPulling="2025-11-25 10:41:08.993604491 +0000 UTC m=+265.807221078" observedRunningTime="2025-11-25 10:41:10.603937098 +0000 UTC m=+267.417553715" watchObservedRunningTime="2025-11-25 10:41:10.607984667 +0000 UTC m=+267.421601274" Nov 25 10:41:12 crc kubenswrapper[4696]: I1125 10:41:12.656081 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:12 crc kubenswrapper[4696]: I1125 10:41:12.656930 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:12 crc kubenswrapper[4696]: I1125 10:41:12.697808 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:12 crc kubenswrapper[4696]: I1125 10:41:12.866525 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:12 crc kubenswrapper[4696]: I1125 10:41:12.866755 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:12 crc kubenswrapper[4696]: I1125 10:41:12.904760 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:13 crc kubenswrapper[4696]: I1125 10:41:13.640553 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 10:41:13 crc kubenswrapper[4696]: I1125 10:41:13.646871 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h9ssz" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.062321 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.063046 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.102550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.269214 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.269274 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.314856 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.645936 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nbc7f" Nov 25 10:41:15 crc kubenswrapper[4696]: I1125 10:41:15.648003 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 10:43:00 crc kubenswrapper[4696]: I1125 10:43:00.802466 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:43:00 crc kubenswrapper[4696]: I1125 10:43:00.803053 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:43:30 crc kubenswrapper[4696]: I1125 10:43:30.802366 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:43:30 crc kubenswrapper[4696]: I1125 10:43:30.802938 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.425213 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nn5p4"] Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.426475 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.445703 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nn5p4"] Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.482633 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.482710 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e4a97be-5a6e-4715-b989-2bd932b028a3-trusted-ca\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.482730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrnrn\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-kube-api-access-qrnrn\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.482759 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-bound-sa-token\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.482948 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2e4a97be-5a6e-4715-b989-2bd932b028a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.483046 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-registry-tls\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.483227 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2e4a97be-5a6e-4715-b989-2bd932b028a3-registry-certificates\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.483342 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2e4a97be-5a6e-4715-b989-2bd932b028a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.512339 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584216 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e4a97be-5a6e-4715-b989-2bd932b028a3-trusted-ca\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584261 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrnrn\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-kube-api-access-qrnrn\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584292 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-bound-sa-token\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584332 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2e4a97be-5a6e-4715-b989-2bd932b028a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584355 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-registry-tls\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584387 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2e4a97be-5a6e-4715-b989-2bd932b028a3-registry-certificates\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.584421 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2e4a97be-5a6e-4715-b989-2bd932b028a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.585106 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2e4a97be-5a6e-4715-b989-2bd932b028a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.586073 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2e4a97be-5a6e-4715-b989-2bd932b028a3-registry-certificates\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.586138 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e4a97be-5a6e-4715-b989-2bd932b028a3-trusted-ca\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.595449 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-registry-tls\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.595499 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2e4a97be-5a6e-4715-b989-2bd932b028a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.602395 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrnrn\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-kube-api-access-qrnrn\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.613156 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e4a97be-5a6e-4715-b989-2bd932b028a3-bound-sa-token\") pod \"image-registry-66df7c8f76-nn5p4\" (UID: \"2e4a97be-5a6e-4715-b989-2bd932b028a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.746752 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:54 crc kubenswrapper[4696]: I1125 10:43:54.977527 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nn5p4"] Nov 25 10:43:55 crc kubenswrapper[4696]: I1125 10:43:55.585523 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" event={"ID":"2e4a97be-5a6e-4715-b989-2bd932b028a3","Type":"ContainerStarted","Data":"8e43dae19d532a2ffc9712b01fadd1b276485635795a5663ebdcfa0ed4fd07ee"} Nov 25 10:43:55 crc kubenswrapper[4696]: I1125 10:43:55.585578 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" event={"ID":"2e4a97be-5a6e-4715-b989-2bd932b028a3","Type":"ContainerStarted","Data":"e9f9f19faa8103e7a011810161b1ea93a017110dafd7870800fab484b929ce39"} Nov 25 10:43:55 crc kubenswrapper[4696]: I1125 10:43:55.585732 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:43:55 crc kubenswrapper[4696]: I1125 10:43:55.610789 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" podStartSLOduration=1.610761428 podStartE2EDuration="1.610761428s" podCreationTimestamp="2025-11-25 10:43:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:43:55.607214943 +0000 UTC m=+432.420831530" watchObservedRunningTime="2025-11-25 10:43:55.610761428 +0000 UTC m=+432.424378015" Nov 25 10:44:00 crc kubenswrapper[4696]: I1125 10:44:00.802141 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:44:00 crc kubenswrapper[4696]: I1125 10:44:00.802555 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:44:00 crc kubenswrapper[4696]: I1125 10:44:00.802617 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:44:00 crc kubenswrapper[4696]: I1125 10:44:00.803526 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1fb36885e86a8194f5415b7d52233333dec593850e5cf8d383f363bb7a6ad19c"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:44:00 crc kubenswrapper[4696]: I1125 10:44:00.803629 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://1fb36885e86a8194f5415b7d52233333dec593850e5cf8d383f363bb7a6ad19c" gracePeriod=600 Nov 25 10:44:01 crc kubenswrapper[4696]: I1125 10:44:01.627117 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="1fb36885e86a8194f5415b7d52233333dec593850e5cf8d383f363bb7a6ad19c" exitCode=0 Nov 25 10:44:01 crc kubenswrapper[4696]: I1125 10:44:01.627177 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"1fb36885e86a8194f5415b7d52233333dec593850e5cf8d383f363bb7a6ad19c"} Nov 25 10:44:01 crc kubenswrapper[4696]: I1125 10:44:01.627217 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"e613ae8723bcb59d28226fb02f66886a567bae947befcc699ab9dd24e0648204"} Nov 25 10:44:01 crc kubenswrapper[4696]: I1125 10:44:01.627243 4696 scope.go:117] "RemoveContainer" containerID="515c4c74b51d84a02d7b7d4e35cd63e88d93b5b8809f2999815002ff982bc051" Nov 25 10:44:14 crc kubenswrapper[4696]: I1125 10:44:14.756092 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nn5p4" Nov 25 10:44:14 crc kubenswrapper[4696]: I1125 10:44:14.831141 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xkw2f"] Nov 25 10:44:39 crc kubenswrapper[4696]: I1125 10:44:39.884332 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" podUID="545452b4-d392-4d46-8e5e-a2a4c9e7b43b" containerName="registry" containerID="cri-o://19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362" gracePeriod=30 Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.240572 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.324769 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-ca-trust-extracted\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.324856 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29w79\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-kube-api-access-29w79\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.324881 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-trusted-ca\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.324907 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-tls\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.324948 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-installation-pull-secrets\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.325018 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-certificates\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.325154 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.325186 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-bound-sa-token\") pod \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\" (UID: \"545452b4-d392-4d46-8e5e-a2a4c9e7b43b\") " Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.326258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.326300 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.330912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.331107 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.331454 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.332008 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-kube-api-access-29w79" (OuterVolumeSpecName: "kube-api-access-29w79") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "kube-api-access-29w79". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.343890 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.346047 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "545452b4-d392-4d46-8e5e-a2a4c9e7b43b" (UID: "545452b4-d392-4d46-8e5e-a2a4c9e7b43b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.426561 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.426797 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29w79\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-kube-api-access-29w79\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.426883 4696 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.426945 4696 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.426994 4696 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.427043 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.427099 4696 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/545452b4-d392-4d46-8e5e-a2a4c9e7b43b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.874588 4696 generic.go:334] "Generic (PLEG): container finished" podID="545452b4-d392-4d46-8e5e-a2a4c9e7b43b" containerID="19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362" exitCode=0 Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.874653 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" event={"ID":"545452b4-d392-4d46-8e5e-a2a4c9e7b43b","Type":"ContainerDied","Data":"19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362"} Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.874758 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" event={"ID":"545452b4-d392-4d46-8e5e-a2a4c9e7b43b","Type":"ContainerDied","Data":"349ba1afc3174da7195e9fe5015cf8c45a83e90387e12f396583a89fe0a90f77"} Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.874789 4696 scope.go:117] "RemoveContainer" containerID="19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.875268 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xkw2f" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.912621 4696 scope.go:117] "RemoveContainer" containerID="19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362" Nov 25 10:44:40 crc kubenswrapper[4696]: E1125 10:44:40.913370 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362\": container with ID starting with 19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362 not found: ID does not exist" containerID="19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.913442 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362"} err="failed to get container status \"19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362\": rpc error: code = NotFound desc = could not find container \"19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362\": container with ID starting with 19fea422e5f24fc1c628a002a2bb93b55037cba45d2df26f4ee4253521ca8362 not found: ID does not exist" Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.926317 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xkw2f"] Nov 25 10:44:40 crc kubenswrapper[4696]: I1125 10:44:40.933232 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xkw2f"] Nov 25 10:44:42 crc kubenswrapper[4696]: I1125 10:44:42.052920 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="545452b4-d392-4d46-8e5e-a2a4c9e7b43b" path="/var/lib/kubelet/pods/545452b4-d392-4d46-8e5e-a2a4c9e7b43b/volumes" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.872539 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 10:44:43 crc kubenswrapper[4696]: E1125 10:44:43.873314 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="545452b4-d392-4d46-8e5e-a2a4c9e7b43b" containerName="registry" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.873344 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="545452b4-d392-4d46-8e5e-a2a4c9e7b43b" containerName="registry" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.873606 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="545452b4-d392-4d46-8e5e-a2a4c9e7b43b" containerName="registry" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.874402 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.877072 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.877923 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.885550 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.971053 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/461eff7f-160c-4410-aa9c-3b0065cda89c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:43 crc kubenswrapper[4696]: I1125 10:44:43.971105 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/461eff7f-160c-4410-aa9c-3b0065cda89c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.072183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/461eff7f-160c-4410-aa9c-3b0065cda89c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.072260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/461eff7f-160c-4410-aa9c-3b0065cda89c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.072349 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/461eff7f-160c-4410-aa9c-3b0065cda89c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.086657 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.103169 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/461eff7f-160c-4410-aa9c-3b0065cda89c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.200458 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.208259 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.399602 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.898102 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"461eff7f-160c-4410-aa9c-3b0065cda89c","Type":"ContainerStarted","Data":"9f4a9858cd0ae687edac9cc298383a74347357494ce5e0cdb1599c089fd5907b"} Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.900110 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"461eff7f-160c-4410-aa9c-3b0065cda89c","Type":"ContainerStarted","Data":"3b499d0a0ab1315651a978025af6823c7073257cc08eae4f3fabc66504650f06"} Nov 25 10:44:44 crc kubenswrapper[4696]: I1125 10:44:44.913135 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.913096476 podStartE2EDuration="1.913096476s" podCreationTimestamp="2025-11-25 10:44:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:44:44.910912297 +0000 UTC m=+481.724528894" watchObservedRunningTime="2025-11-25 10:44:44.913096476 +0000 UTC m=+481.726713073" Nov 25 10:44:45 crc kubenswrapper[4696]: I1125 10:44:45.908903 4696 generic.go:334] "Generic (PLEG): container finished" podID="461eff7f-160c-4410-aa9c-3b0065cda89c" containerID="9f4a9858cd0ae687edac9cc298383a74347357494ce5e0cdb1599c089fd5907b" exitCode=0 Nov 25 10:44:45 crc kubenswrapper[4696]: I1125 10:44:45.909007 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"461eff7f-160c-4410-aa9c-3b0065cda89c","Type":"ContainerDied","Data":"9f4a9858cd0ae687edac9cc298383a74347357494ce5e0cdb1599c089fd5907b"} Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.104432 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.211373 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/461eff7f-160c-4410-aa9c-3b0065cda89c-kubelet-dir\") pod \"461eff7f-160c-4410-aa9c-3b0065cda89c\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.211486 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/461eff7f-160c-4410-aa9c-3b0065cda89c-kube-api-access\") pod \"461eff7f-160c-4410-aa9c-3b0065cda89c\" (UID: \"461eff7f-160c-4410-aa9c-3b0065cda89c\") " Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.211599 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/461eff7f-160c-4410-aa9c-3b0065cda89c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "461eff7f-160c-4410-aa9c-3b0065cda89c" (UID: "461eff7f-160c-4410-aa9c-3b0065cda89c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.211713 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/461eff7f-160c-4410-aa9c-3b0065cda89c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.216228 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/461eff7f-160c-4410-aa9c-3b0065cda89c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "461eff7f-160c-4410-aa9c-3b0065cda89c" (UID: "461eff7f-160c-4410-aa9c-3b0065cda89c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.312320 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/461eff7f-160c-4410-aa9c-3b0065cda89c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.922751 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"461eff7f-160c-4410-aa9c-3b0065cda89c","Type":"ContainerDied","Data":"3b499d0a0ab1315651a978025af6823c7073257cc08eae4f3fabc66504650f06"} Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.922799 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b499d0a0ab1315651a978025af6823c7073257cc08eae4f3fabc66504650f06" Nov 25 10:44:47 crc kubenswrapper[4696]: I1125 10:44:47.922844 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.272231 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 10:44:50 crc kubenswrapper[4696]: E1125 10:44:50.273082 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461eff7f-160c-4410-aa9c-3b0065cda89c" containerName="pruner" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.273115 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="461eff7f-160c-4410-aa9c-3b0065cda89c" containerName="pruner" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.273396 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="461eff7f-160c-4410-aa9c-3b0065cda89c" containerName="pruner" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.274234 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.277255 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.279286 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.285379 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.353050 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-var-lock\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.353144 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-kubelet-dir\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.353171 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/962cb416-4905-4b0f-a88e-cfd65cfebf69-kube-api-access\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.453959 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-var-lock\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.454046 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-kubelet-dir\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.454074 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/962cb416-4905-4b0f-a88e-cfd65cfebf69-kube-api-access\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.454116 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-kubelet-dir\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.454224 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-var-lock\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.487934 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/962cb416-4905-4b0f-a88e-cfd65cfebf69-kube-api-access\") pod \"installer-9-crc\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.611733 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.825321 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 10:44:50 crc kubenswrapper[4696]: I1125 10:44:50.940153 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"962cb416-4905-4b0f-a88e-cfd65cfebf69","Type":"ContainerStarted","Data":"f351bcf094831e3d392f882473ab7ee38ba0c767d8c583200be1c9812c5efce8"} Nov 25 10:44:51 crc kubenswrapper[4696]: I1125 10:44:51.972301 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"962cb416-4905-4b0f-a88e-cfd65cfebf69","Type":"ContainerStarted","Data":"762cc88765a8056c4ff98d12106c1c355d8eb32d7b7a8083886c33713927900e"} Nov 25 10:44:51 crc kubenswrapper[4696]: I1125 10:44:51.999705 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.9996889420000001 podStartE2EDuration="1.999688942s" podCreationTimestamp="2025-11-25 10:44:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:44:51.997970623 +0000 UTC m=+488.811587260" watchObservedRunningTime="2025-11-25 10:44:51.999688942 +0000 UTC m=+488.813305529" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.149953 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4"] Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.151604 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.153944 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.154049 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.174038 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4"] Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.289387 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-config-volume\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.289455 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2dhd\" (UniqueName: \"kubernetes.io/projected/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-kube-api-access-d2dhd\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.289491 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-secret-volume\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.390021 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-config-volume\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.390066 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2dhd\" (UniqueName: \"kubernetes.io/projected/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-kube-api-access-d2dhd\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.390441 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-secret-volume\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.391324 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-config-volume\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.399157 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-secret-volume\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.409506 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2dhd\" (UniqueName: \"kubernetes.io/projected/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-kube-api-access-d2dhd\") pod \"collect-profiles-29401125-hldk4\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.479814 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:00 crc kubenswrapper[4696]: I1125 10:45:00.655395 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4"] Nov 25 10:45:01 crc kubenswrapper[4696]: I1125 10:45:01.036019 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" event={"ID":"59f046cd-d8e3-4ee3-aa07-2ea312ccf224","Type":"ContainerStarted","Data":"0c35d6b79f6ab2ca262279badacf8b0bae6e8d68bc002fe88eb656634e5cecb8"} Nov 25 10:45:01 crc kubenswrapper[4696]: I1125 10:45:01.036355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" event={"ID":"59f046cd-d8e3-4ee3-aa07-2ea312ccf224","Type":"ContainerStarted","Data":"62be12a1295b7faad7f2e35c07ca4f62e753cb679074c2523b4535941539f487"} Nov 25 10:45:01 crc kubenswrapper[4696]: I1125 10:45:01.056738 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" podStartSLOduration=1.056706016 podStartE2EDuration="1.056706016s" podCreationTimestamp="2025-11-25 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:45:01.053805884 +0000 UTC m=+497.867422481" watchObservedRunningTime="2025-11-25 10:45:01.056706016 +0000 UTC m=+497.870322693" Nov 25 10:45:02 crc kubenswrapper[4696]: I1125 10:45:02.043490 4696 generic.go:334] "Generic (PLEG): container finished" podID="59f046cd-d8e3-4ee3-aa07-2ea312ccf224" containerID="0c35d6b79f6ab2ca262279badacf8b0bae6e8d68bc002fe88eb656634e5cecb8" exitCode=0 Nov 25 10:45:02 crc kubenswrapper[4696]: I1125 10:45:02.049707 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" event={"ID":"59f046cd-d8e3-4ee3-aa07-2ea312ccf224","Type":"ContainerDied","Data":"0c35d6b79f6ab2ca262279badacf8b0bae6e8d68bc002fe88eb656634e5cecb8"} Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.254322 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.323619 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-secret-volume\") pod \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.323656 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2dhd\" (UniqueName: \"kubernetes.io/projected/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-kube-api-access-d2dhd\") pod \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.323701 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-config-volume\") pod \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\" (UID: \"59f046cd-d8e3-4ee3-aa07-2ea312ccf224\") " Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.324524 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-config-volume" (OuterVolumeSpecName: "config-volume") pod "59f046cd-d8e3-4ee3-aa07-2ea312ccf224" (UID: "59f046cd-d8e3-4ee3-aa07-2ea312ccf224"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.328385 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-kube-api-access-d2dhd" (OuterVolumeSpecName: "kube-api-access-d2dhd") pod "59f046cd-d8e3-4ee3-aa07-2ea312ccf224" (UID: "59f046cd-d8e3-4ee3-aa07-2ea312ccf224"). InnerVolumeSpecName "kube-api-access-d2dhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.328401 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59f046cd-d8e3-4ee3-aa07-2ea312ccf224" (UID: "59f046cd-d8e3-4ee3-aa07-2ea312ccf224"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.425230 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.425418 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2dhd\" (UniqueName: \"kubernetes.io/projected/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-kube-api-access-d2dhd\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:03 crc kubenswrapper[4696]: I1125 10:45:03.425501 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59f046cd-d8e3-4ee3-aa07-2ea312ccf224-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:04 crc kubenswrapper[4696]: I1125 10:45:04.058974 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" Nov 25 10:45:04 crc kubenswrapper[4696]: I1125 10:45:04.062271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4" event={"ID":"59f046cd-d8e3-4ee3-aa07-2ea312ccf224","Type":"ContainerDied","Data":"62be12a1295b7faad7f2e35c07ca4f62e753cb679074c2523b4535941539f487"} Nov 25 10:45:04 crc kubenswrapper[4696]: I1125 10:45:04.062316 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62be12a1295b7faad7f2e35c07ca4f62e753cb679074c2523b4535941539f487" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.038475 4696 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.039268 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f046cd-d8e3-4ee3-aa07-2ea312ccf224" containerName="collect-profiles" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.039287 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f046cd-d8e3-4ee3-aa07-2ea312ccf224" containerName="collect-profiles" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.040952 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f046cd-d8e3-4ee3-aa07-2ea312ccf224" containerName="collect-profiles" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.041541 4696 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.041820 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.042062 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478" gracePeriod=15 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.042220 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042" gracePeriod=15 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.042309 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643" gracePeriod=15 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.042289 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b" gracePeriod=15 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.042474 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3" gracePeriod=15 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.059012 4696 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.061879 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.061908 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.061921 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.061930 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.061957 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.061965 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.061978 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062009 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.062040 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062049 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.062071 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062080 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.062098 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062107 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.062124 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062133 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062435 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062457 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062472 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062483 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062501 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062516 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.062960 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.100295 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143193 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143255 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143281 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143327 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143376 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143422 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143447 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.143468 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.188899 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.190547 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.191601 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042" exitCode=0 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.191632 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3" exitCode=0 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.191643 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b" exitCode=0 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.191651 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643" exitCode=2 Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.191707 4696 scope.go:117] "RemoveContainer" containerID="9545251362ef081c3b972973a74d8ca70be87faa075b6a66e7fe10a16dfdcc4f" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244565 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244610 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244655 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244680 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244722 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244746 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244786 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244803 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244805 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244822 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244843 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244887 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244846 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244930 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.244985 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.389532 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:45:29 crc kubenswrapper[4696]: E1125 10:45:29.417150 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.38:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b3a1252a79fcc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:45:29.41622062 +0000 UTC m=+526.229837247,LastTimestamp:2025-11-25 10:45:29.41622062 +0000 UTC m=+526.229837247,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.617127 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 10:45:29 crc kubenswrapper[4696]: I1125 10:45:29.617437 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.198883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"8fbfa8a016c353e9f897e2b4b5437e7f0aa8e1fee37c614360365636553184dd"} Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.198960 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"47f090658e8167186d24f30d40b5feac88218dbbef82c351ea47996eeb736d03"} Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.200611 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.202026 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.204629 4696 generic.go:334] "Generic (PLEG): container finished" podID="962cb416-4905-4b0f-a88e-cfd65cfebf69" containerID="762cc88765a8056c4ff98d12106c1c355d8eb32d7b7a8083886c33713927900e" exitCode=0 Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.204676 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"962cb416-4905-4b0f-a88e-cfd65cfebf69","Type":"ContainerDied","Data":"762cc88765a8056c4ff98d12106c1c355d8eb32d7b7a8083886c33713927900e"} Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.205131 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: I1125 10:45:30.205589 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: E1125 10:45:30.580855 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:30Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:30Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:30Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:30Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: E1125 10:45:30.581265 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: E1125 10:45:30.581732 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: E1125 10:45:30.582307 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: E1125 10:45:30.582756 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:30 crc kubenswrapper[4696]: E1125 10:45:30.582801 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.419103 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.420590 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.421154 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.421529 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.421807 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.457072 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.457593 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.458058 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.458392 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.475822 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.475874 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.475902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.475968 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.475984 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.476061 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.476196 4696 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.476211 4696 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.476222 4696 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577246 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/962cb416-4905-4b0f-a88e-cfd65cfebf69-kube-api-access\") pod \"962cb416-4905-4b0f-a88e-cfd65cfebf69\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577298 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-var-lock\") pod \"962cb416-4905-4b0f-a88e-cfd65cfebf69\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577350 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-kubelet-dir\") pod \"962cb416-4905-4b0f-a88e-cfd65cfebf69\" (UID: \"962cb416-4905-4b0f-a88e-cfd65cfebf69\") " Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577534 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-var-lock" (OuterVolumeSpecName: "var-lock") pod "962cb416-4905-4b0f-a88e-cfd65cfebf69" (UID: "962cb416-4905-4b0f-a88e-cfd65cfebf69"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577559 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "962cb416-4905-4b0f-a88e-cfd65cfebf69" (UID: "962cb416-4905-4b0f-a88e-cfd65cfebf69"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577696 4696 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.577712 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/962cb416-4905-4b0f-a88e-cfd65cfebf69-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.584603 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962cb416-4905-4b0f-a88e-cfd65cfebf69-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "962cb416-4905-4b0f-a88e-cfd65cfebf69" (UID: "962cb416-4905-4b0f-a88e-cfd65cfebf69"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:45:31 crc kubenswrapper[4696]: I1125 10:45:31.679149 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/962cb416-4905-4b0f-a88e-cfd65cfebf69-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.050023 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.224645 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.226402 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478" exitCode=0 Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.226543 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.226566 4696 scope.go:117] "RemoveContainer" containerID="ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.227192 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.229101 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.229492 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.230783 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"962cb416-4905-4b0f-a88e-cfd65cfebf69","Type":"ContainerDied","Data":"f351bcf094831e3d392f882473ab7ee38ba0c767d8c583200be1c9812c5efce8"} Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.230819 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f351bcf094831e3d392f882473ab7ee38ba0c767d8c583200be1c9812c5efce8" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.230903 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.231195 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.231481 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.232854 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.238153 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.238888 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.239229 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.253145 4696 scope.go:117] "RemoveContainer" containerID="527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.273879 4696 scope.go:117] "RemoveContainer" containerID="fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.293741 4696 scope.go:117] "RemoveContainer" containerID="e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.315948 4696 scope.go:117] "RemoveContainer" containerID="8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.351315 4696 scope.go:117] "RemoveContainer" containerID="af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.378615 4696 scope.go:117] "RemoveContainer" containerID="ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042" Nov 25 10:45:32 crc kubenswrapper[4696]: E1125 10:45:32.379740 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\": container with ID starting with ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042 not found: ID does not exist" containerID="ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.379868 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042"} err="failed to get container status \"ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\": rpc error: code = NotFound desc = could not find container \"ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042\": container with ID starting with ac55aeae117b2d2ff3d2a2b694f9b84ca8690916002c248f735be97bd483f042 not found: ID does not exist" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.379930 4696 scope.go:117] "RemoveContainer" containerID="527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3" Nov 25 10:45:32 crc kubenswrapper[4696]: E1125 10:45:32.380530 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\": container with ID starting with 527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3 not found: ID does not exist" containerID="527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.380617 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3"} err="failed to get container status \"527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\": rpc error: code = NotFound desc = could not find container \"527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3\": container with ID starting with 527af5f29aa1e1e95685c9e85a040e00fd83ddf88de6e467761ec24e2c9532b3 not found: ID does not exist" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.380686 4696 scope.go:117] "RemoveContainer" containerID="fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b" Nov 25 10:45:32 crc kubenswrapper[4696]: E1125 10:45:32.381986 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\": container with ID starting with fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b not found: ID does not exist" containerID="fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.382018 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b"} err="failed to get container status \"fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\": rpc error: code = NotFound desc = could not find container \"fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b\": container with ID starting with fdde4a79dcaf5ee9548c92a4d14d2a66d0611b1bb3d0040bed92804a3535596b not found: ID does not exist" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.382061 4696 scope.go:117] "RemoveContainer" containerID="e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643" Nov 25 10:45:32 crc kubenswrapper[4696]: E1125 10:45:32.382436 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\": container with ID starting with e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643 not found: ID does not exist" containerID="e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.382496 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643"} err="failed to get container status \"e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\": rpc error: code = NotFound desc = could not find container \"e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643\": container with ID starting with e67d19a7a33433e2dd11a08088d0ed2e3eceac6938f83cbb12c22aa902673643 not found: ID does not exist" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.382512 4696 scope.go:117] "RemoveContainer" containerID="8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478" Nov 25 10:45:32 crc kubenswrapper[4696]: E1125 10:45:32.382939 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\": container with ID starting with 8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478 not found: ID does not exist" containerID="8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.382983 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478"} err="failed to get container status \"8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\": rpc error: code = NotFound desc = could not find container \"8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478\": container with ID starting with 8f25476e85b13e1a2f08bda1ad4ccbc9fcf6ddbc3c1042a7ee752c90e3bc8478 not found: ID does not exist" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.383012 4696 scope.go:117] "RemoveContainer" containerID="af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e" Nov 25 10:45:32 crc kubenswrapper[4696]: E1125 10:45:32.383429 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\": container with ID starting with af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e not found: ID does not exist" containerID="af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e" Nov 25 10:45:32 crc kubenswrapper[4696]: I1125 10:45:32.383554 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e"} err="failed to get container status \"af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\": rpc error: code = NotFound desc = could not find container \"af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e\": container with ID starting with af94d4e73d06bad5cf2ff0400b593c8f256f54cae6b4a3d72250a780a7a20f2e not found: ID does not exist" Nov 25 10:45:34 crc kubenswrapper[4696]: I1125 10:45:34.044263 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:34 crc kubenswrapper[4696]: I1125 10:45:34.044633 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:34 crc kubenswrapper[4696]: I1125 10:45:34.044896 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:36 crc kubenswrapper[4696]: E1125 10:45:36.354178 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.38:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b3a1252a79fcc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:45:29.41622062 +0000 UTC m=+526.229837247,LastTimestamp:2025-11-25 10:45:29.41622062 +0000 UTC m=+526.229837247,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.594924 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.595623 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.596301 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.596795 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.597263 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:38 crc kubenswrapper[4696]: I1125 10:45:38.597322 4696 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.597985 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="200ms" Nov 25 10:45:38 crc kubenswrapper[4696]: E1125 10:45:38.799272 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="400ms" Nov 25 10:45:39 crc kubenswrapper[4696]: E1125 10:45:39.200728 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="800ms" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.002068 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="1.6s" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.987210 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:40Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:40Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:40Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:45:40Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.987923 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.988165 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.988411 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.988788 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:40 crc kubenswrapper[4696]: E1125 10:45:40.988831 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:45:41 crc kubenswrapper[4696]: E1125 10:45:41.603626 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.38:6443: connect: connection refused" interval="3.2s" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.041900 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.042978 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.043479 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.056147 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.056191 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:42 crc kubenswrapper[4696]: E1125 10:45:42.056733 4696 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.057495 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:42 crc kubenswrapper[4696]: W1125 10:45:42.077863 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-3df658487562af95a410f50a0be13eb760b7d271c4a5fb2cc23a3246b70e715c WatchSource:0}: Error finding container 3df658487562af95a410f50a0be13eb760b7d271c4a5fb2cc23a3246b70e715c: Status 404 returned error can't find the container with id 3df658487562af95a410f50a0be13eb760b7d271c4a5fb2cc23a3246b70e715c Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.291416 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.291768 4696 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523" exitCode=1 Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.291845 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523"} Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.292406 4696 scope.go:117] "RemoveContainer" containerID="027992d2c8e97aa045eceb5a4a0b06c618e8dd1916a3adcf14429d3c50624523" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.292748 4696 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.293204 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.293555 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:42 crc kubenswrapper[4696]: I1125 10:45:42.294179 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3df658487562af95a410f50a0be13eb760b7d271c4a5fb2cc23a3246b70e715c"} Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.306442 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.306609 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f1a4df019d73353851b8bb8600932c16e0447547a6a0ec3e2e392c039fda3e0f"} Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.307845 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.308348 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.309004 4696 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.309903 4696 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c8d0f296242d33845f3e4613c4c8ed03d686a5ec5da651b5f39a1639a92a9973" exitCode=0 Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.309991 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c8d0f296242d33845f3e4613c4c8ed03d686a5ec5da651b5f39a1639a92a9973"} Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.310386 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.310424 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.311010 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:43 crc kubenswrapper[4696]: E1125 10:45:43.311111 4696 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.311534 4696 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:43 crc kubenswrapper[4696]: I1125 10:45:43.312139 4696 status_manager.go:851] "Failed to get status for pod" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.38:6443: connect: connection refused" Nov 25 10:45:44 crc kubenswrapper[4696]: I1125 10:45:44.322314 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9fb12e51a59e207a4477e8272de93ac5ec116a45c76bba4ddb8d974ad27823fe"} Nov 25 10:45:44 crc kubenswrapper[4696]: I1125 10:45:44.322361 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e7413508adc565c9103cef9e26bd2278bb16db5c08bda300c9c8215fa4fb3575"} Nov 25 10:45:44 crc kubenswrapper[4696]: I1125 10:45:44.322372 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3eec8d86a709bcc3ae2daaf185a63c3ab529b85b6edb4e28777fdfb1a22bfc33"} Nov 25 10:45:45 crc kubenswrapper[4696]: I1125 10:45:45.330487 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e8bacf957dbb01b7bb3d6f952ca61fc21262286eec0f64a64bac79ad0b06a2e"} Nov 25 10:45:45 crc kubenswrapper[4696]: I1125 10:45:45.330824 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5065b8b9742622d771991da37a5b0cd84743a9aa98f3f5c28a9ba4c9d660992c"} Nov 25 10:45:45 crc kubenswrapper[4696]: I1125 10:45:45.330850 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:45 crc kubenswrapper[4696]: I1125 10:45:45.330768 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:45 crc kubenswrapper[4696]: I1125 10:45:45.330873 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:47 crc kubenswrapper[4696]: I1125 10:45:47.058615 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:47 crc kubenswrapper[4696]: I1125 10:45:47.059704 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:47 crc kubenswrapper[4696]: I1125 10:45:47.063851 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:47 crc kubenswrapper[4696]: I1125 10:45:47.144824 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:45:50 crc kubenswrapper[4696]: I1125 10:45:50.338095 4696 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:51 crc kubenswrapper[4696]: I1125 10:45:51.195881 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:45:51 crc kubenswrapper[4696]: I1125 10:45:51.199760 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:45:51 crc kubenswrapper[4696]: I1125 10:45:51.359865 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:51 crc kubenswrapper[4696]: I1125 10:45:51.359898 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:51 crc kubenswrapper[4696]: I1125 10:45:51.363415 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:45:51 crc kubenswrapper[4696]: I1125 10:45:51.365543 4696 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0408e79e-ea0f-4679-a551-edff9abfeeef" Nov 25 10:45:52 crc kubenswrapper[4696]: I1125 10:45:52.368137 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:52 crc kubenswrapper[4696]: I1125 10:45:52.368178 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="34438097-5858-4502-983c-3e167dea3525" Nov 25 10:45:54 crc kubenswrapper[4696]: I1125 10:45:54.053608 4696 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0408e79e-ea0f-4679-a551-edff9abfeeef" Nov 25 10:45:57 crc kubenswrapper[4696]: I1125 10:45:57.148621 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:46:00 crc kubenswrapper[4696]: I1125 10:46:00.110795 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 10:46:00 crc kubenswrapper[4696]: I1125 10:46:00.122265 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 10:46:00 crc kubenswrapper[4696]: I1125 10:46:00.364025 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 10:46:00 crc kubenswrapper[4696]: I1125 10:46:00.809326 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 10:46:00 crc kubenswrapper[4696]: I1125 10:46:00.931698 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:46:01 crc kubenswrapper[4696]: I1125 10:46:01.073711 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 10:46:01 crc kubenswrapper[4696]: I1125 10:46:01.433803 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.191848 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.285547 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.347795 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.485794 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.531338 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.591218 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.776633 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.891651 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 10:46:02 crc kubenswrapper[4696]: I1125 10:46:02.912240 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.204973 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.341648 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.470437 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.635631 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.724538 4696 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.730304 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.770547 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.772458 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.890707 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 10:46:03 crc kubenswrapper[4696]: I1125 10:46:03.983221 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.145116 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.162260 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.232309 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.251567 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.566141 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.580046 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.582950 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.648148 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.702007 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.782191 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.808509 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.921501 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.941257 4696 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 10:46:04 crc kubenswrapper[4696]: I1125 10:46:04.985210 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.261523 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.402920 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.483023 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.540816 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.655609 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.693613 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.759263 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.817341 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.855966 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.859629 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.901546 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 10:46:05 crc kubenswrapper[4696]: I1125 10:46:05.979729 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.225033 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.316883 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.347949 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.371431 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.469566 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.585257 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.637540 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.698494 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.713007 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.803627 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.831956 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.904725 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.963914 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 10:46:06 crc kubenswrapper[4696]: I1125 10:46:06.965359 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.049542 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.081178 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.104010 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.137490 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.167101 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.226353 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.258536 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.316538 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.328557 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.377076 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.627174 4696 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.700325 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.733903 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.739760 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.892092 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 10:46:07 crc kubenswrapper[4696]: I1125 10:46:07.951868 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.086787 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.137887 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.186941 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.228326 4696 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.237967 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.285358 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.346870 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.390113 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.392795 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.393231 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.471517 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.564191 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.592786 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.663424 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.721790 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.768334 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.780086 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.884348 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.956747 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.959611 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 10:46:08 crc kubenswrapper[4696]: I1125 10:46:08.959727 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.011756 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.040744 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.107399 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.208643 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.278864 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.339010 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.415204 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.473975 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.655142 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.741963 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.750685 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.750792 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.771873 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.788731 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.833422 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 10:46:09 crc kubenswrapper[4696]: I1125 10:46:09.855880 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.016940 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.050354 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.058817 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.169704 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.173775 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.174840 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.176365 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.195867 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.206262 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.314344 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.349122 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.380138 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.384261 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.401742 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.428560 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.441694 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.467739 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.481309 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.533133 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.538372 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.635638 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.635756 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.640855 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.646621 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.654490 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.696809 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.724037 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.726432 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.766112 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.799916 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.824316 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.856039 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.863763 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.893569 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 10:46:10 crc kubenswrapper[4696]: I1125 10:46:10.926639 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.169305 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.205332 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.206112 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.209384 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.300139 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.301947 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.432926 4696 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.434729 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.436055 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.436029161 podStartE2EDuration="42.436029161s" podCreationTimestamp="2025-11-25 10:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:45:50.275325435 +0000 UTC m=+547.088942042" watchObservedRunningTime="2025-11-25 10:46:11.436029161 +0000 UTC m=+568.249645778" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.441473 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.441559 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.447286 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.470588 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.470570443 podStartE2EDuration="21.470570443s" podCreationTimestamp="2025-11-25 10:45:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:11.467441091 +0000 UTC m=+568.281057678" watchObservedRunningTime="2025-11-25 10:46:11.470570443 +0000 UTC m=+568.284187030" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.487325 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.504166 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.506332 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.508041 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.538335 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.562616 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.590498 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.838034 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.842127 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.869144 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.927634 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 10:46:11 crc kubenswrapper[4696]: I1125 10:46:11.975588 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.050902 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.088056 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.142404 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.157799 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.194259 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.281745 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.298240 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.343484 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.364104 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.402041 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.444105 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.471153 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.513553 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.518196 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.524765 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.707326 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.749063 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.751283 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.793301 4696 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.794478 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.799243 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.820003 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.829857 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.863872 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.970957 4696 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:46:12 crc kubenswrapper[4696]: I1125 10:46:12.971198 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://8fbfa8a016c353e9f897e2b4b5437e7f0aa8e1fee37c614360365636553184dd" gracePeriod=5 Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.056642 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.182598 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.210146 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.240740 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.264046 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.290161 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.307259 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.379917 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.447201 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.471513 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.567995 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.649496 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.710938 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.856200 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.860412 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 10:46:13 crc kubenswrapper[4696]: I1125 10:46:13.951437 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.098826 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.171790 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.191029 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.248573 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.315574 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.337846 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.339811 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.371038 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.374712 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.383713 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.648162 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.648731 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.664658 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.698591 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.709441 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.802010 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.868167 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 10:46:14 crc kubenswrapper[4696]: I1125 10:46:14.924396 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.005927 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.049351 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.075046 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.105410 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.116204 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.125214 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.257006 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.420279 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 10:46:15 crc kubenswrapper[4696]: I1125 10:46:15.469783 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.078355 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.220590 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.486067 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.569945 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.643015 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.671368 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 10:46:16 crc kubenswrapper[4696]: I1125 10:46:16.929938 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 10:46:17 crc kubenswrapper[4696]: I1125 10:46:17.254213 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 10:46:17 crc kubenswrapper[4696]: I1125 10:46:17.448793 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 10:46:17 crc kubenswrapper[4696]: I1125 10:46:17.513057 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 10:46:17 crc kubenswrapper[4696]: I1125 10:46:17.897727 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.515529 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.515821 4696 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="8fbfa8a016c353e9f897e2b4b5437e7f0aa8e1fee37c614360365636553184dd" exitCode=137 Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.689419 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.689499 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793034 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793114 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793181 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793375 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793423 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793446 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793484 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793605 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793751 4696 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793773 4696 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793786 4696 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.793554 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.804979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.895021 4696 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:18 crc kubenswrapper[4696]: I1125 10:46:18.895073 4696 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:19 crc kubenswrapper[4696]: I1125 10:46:19.523847 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 10:46:19 crc kubenswrapper[4696]: I1125 10:46:19.523922 4696 scope.go:117] "RemoveContainer" containerID="8fbfa8a016c353e9f897e2b4b5437e7f0aa8e1fee37c614360365636553184dd" Nov 25 10:46:19 crc kubenswrapper[4696]: I1125 10:46:19.523985 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:46:20 crc kubenswrapper[4696]: I1125 10:46:20.053371 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 10:46:20 crc kubenswrapper[4696]: I1125 10:46:20.053835 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 25 10:46:20 crc kubenswrapper[4696]: I1125 10:46:20.066490 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:46:20 crc kubenswrapper[4696]: I1125 10:46:20.066519 4696 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="50055b78-8129-446c-84e3-d195eccd2c4a" Nov 25 10:46:20 crc kubenswrapper[4696]: I1125 10:46:20.076154 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:46:20 crc kubenswrapper[4696]: I1125 10:46:20.076205 4696 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="50055b78-8129-446c-84e3-d195eccd2c4a" Nov 25 10:46:30 crc kubenswrapper[4696]: I1125 10:46:30.802325 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:46:30 crc kubenswrapper[4696]: I1125 10:46:30.803097 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:46:31 crc kubenswrapper[4696]: I1125 10:46:31.595794 4696 generic.go:334] "Generic (PLEG): container finished" podID="e77b3a6e-c1e9-44ef-806f-6adcba39fe31" containerID="38b63ce85fbb6cbc7fc753cb0cebb9ab50b7c027ea18b36395d9bba487f5a695" exitCode=0 Nov 25 10:46:31 crc kubenswrapper[4696]: I1125 10:46:31.595903 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" event={"ID":"e77b3a6e-c1e9-44ef-806f-6adcba39fe31","Type":"ContainerDied","Data":"38b63ce85fbb6cbc7fc753cb0cebb9ab50b7c027ea18b36395d9bba487f5a695"} Nov 25 10:46:31 crc kubenswrapper[4696]: I1125 10:46:31.596904 4696 scope.go:117] "RemoveContainer" containerID="38b63ce85fbb6cbc7fc753cb0cebb9ab50b7c027ea18b36395d9bba487f5a695" Nov 25 10:46:32 crc kubenswrapper[4696]: I1125 10:46:32.602408 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" event={"ID":"e77b3a6e-c1e9-44ef-806f-6adcba39fe31","Type":"ContainerStarted","Data":"3634306cfc1e7c349d215905963982070d87f506ba94408f74b2c3b31ce1edf5"} Nov 25 10:46:32 crc kubenswrapper[4696]: I1125 10:46:32.603586 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:46:32 crc kubenswrapper[4696]: I1125 10:46:32.604186 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xrn8c" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.956567 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qt985"] Nov 25 10:46:44 crc kubenswrapper[4696]: E1125 10:46:44.957036 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.957048 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 10:46:44 crc kubenswrapper[4696]: E1125 10:46:44.957063 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" containerName="installer" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.957070 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" containerName="installer" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.957978 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="962cb416-4905-4b0f-a88e-cfd65cfebf69" containerName="installer" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.957996 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.958383 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.960232 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.960264 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.968845 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qt985"] Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.990719 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-knqmh"] Nov 25 10:46:44 crc kubenswrapper[4696]: I1125 10:46:44.991538 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.005041 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dvr5m"] Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.005944 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dvr5m" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.008953 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-knqmh"] Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.022683 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dvr5m"] Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.122344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb72g\" (UniqueName: \"kubernetes.io/projected/ed104e46-ddf7-496c-a389-1c78d24e8e7c-kube-api-access-bb72g\") pod \"cert-manager-webhook-5655c58dd6-knqmh\" (UID: \"ed104e46-ddf7-496c-a389-1c78d24e8e7c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.122427 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hl4m\" (UniqueName: \"kubernetes.io/projected/c4daf7cb-c6fd-43da-8821-d26fe9f88fac-kube-api-access-4hl4m\") pod \"cert-manager-5b446d88c5-dvr5m\" (UID: \"c4daf7cb-c6fd-43da-8821-d26fe9f88fac\") " pod="cert-manager/cert-manager-5b446d88c5-dvr5m" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.122473 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcrqm\" (UniqueName: \"kubernetes.io/projected/c1f44937-600c-4df9-94f4-8adbc7a9b0ef-kube-api-access-dcrqm\") pod \"cert-manager-cainjector-7f985d654d-qt985\" (UID: \"c1f44937-600c-4df9-94f4-8adbc7a9b0ef\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.223983 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb72g\" (UniqueName: \"kubernetes.io/projected/ed104e46-ddf7-496c-a389-1c78d24e8e7c-kube-api-access-bb72g\") pod \"cert-manager-webhook-5655c58dd6-knqmh\" (UID: \"ed104e46-ddf7-496c-a389-1c78d24e8e7c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.224040 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hl4m\" (UniqueName: \"kubernetes.io/projected/c4daf7cb-c6fd-43da-8821-d26fe9f88fac-kube-api-access-4hl4m\") pod \"cert-manager-5b446d88c5-dvr5m\" (UID: \"c4daf7cb-c6fd-43da-8821-d26fe9f88fac\") " pod="cert-manager/cert-manager-5b446d88c5-dvr5m" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.224083 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcrqm\" (UniqueName: \"kubernetes.io/projected/c1f44937-600c-4df9-94f4-8adbc7a9b0ef-kube-api-access-dcrqm\") pod \"cert-manager-cainjector-7f985d654d-qt985\" (UID: \"c1f44937-600c-4df9-94f4-8adbc7a9b0ef\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.254079 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb72g\" (UniqueName: \"kubernetes.io/projected/ed104e46-ddf7-496c-a389-1c78d24e8e7c-kube-api-access-bb72g\") pod \"cert-manager-webhook-5655c58dd6-knqmh\" (UID: \"ed104e46-ddf7-496c-a389-1c78d24e8e7c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.262612 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hl4m\" (UniqueName: \"kubernetes.io/projected/c4daf7cb-c6fd-43da-8821-d26fe9f88fac-kube-api-access-4hl4m\") pod \"cert-manager-5b446d88c5-dvr5m\" (UID: \"c4daf7cb-c6fd-43da-8821-d26fe9f88fac\") " pod="cert-manager/cert-manager-5b446d88c5-dvr5m" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.265779 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcrqm\" (UniqueName: \"kubernetes.io/projected/c1f44937-600c-4df9-94f4-8adbc7a9b0ef-kube-api-access-dcrqm\") pod \"cert-manager-cainjector-7f985d654d-qt985\" (UID: \"c1f44937-600c-4df9-94f4-8adbc7a9b0ef\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.276633 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.305431 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.319522 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dvr5m" Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.529392 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qt985"] Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.538705 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.680840 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" event={"ID":"c1f44937-600c-4df9-94f4-8adbc7a9b0ef","Type":"ContainerStarted","Data":"b948d3bbe967869f7fd1c99fc3484d27303a707d9b11f05a3b782dba1cc54d6e"} Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.786342 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dvr5m"] Nov 25 10:46:45 crc kubenswrapper[4696]: I1125 10:46:45.790706 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-knqmh"] Nov 25 10:46:46 crc kubenswrapper[4696]: I1125 10:46:46.689314 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" event={"ID":"ed104e46-ddf7-496c-a389-1c78d24e8e7c","Type":"ContainerStarted","Data":"a23a02dba66dd0b80059de9cf94ad7b06e197482499480f03e7796241939064f"} Nov 25 10:46:46 crc kubenswrapper[4696]: I1125 10:46:46.690824 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dvr5m" event={"ID":"c4daf7cb-c6fd-43da-8821-d26fe9f88fac","Type":"ContainerStarted","Data":"59873e2c164b915478e4141b1f401237a55f3b8c91ac1b9b62bfbac4bb7ac19a"} Nov 25 10:46:49 crc kubenswrapper[4696]: I1125 10:46:49.711941 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dvr5m" event={"ID":"c4daf7cb-c6fd-43da-8821-d26fe9f88fac","Type":"ContainerStarted","Data":"bedc224fe52b8bde81239fd147ab15accfc38f47bd30db8d9d27953e3d3f82b5"} Nov 25 10:46:49 crc kubenswrapper[4696]: I1125 10:46:49.716562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" event={"ID":"c1f44937-600c-4df9-94f4-8adbc7a9b0ef","Type":"ContainerStarted","Data":"3f7f43e7917d583572b3c6108a966ad901b1ba6f1b7bcbdd98a8407a2b3de906"} Nov 25 10:46:49 crc kubenswrapper[4696]: I1125 10:46:49.738074 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-dvr5m" podStartSLOduration=2.796106124 podStartE2EDuration="5.738044148s" podCreationTimestamp="2025-11-25 10:46:44 +0000 UTC" firstStartedPulling="2025-11-25 10:46:45.796409879 +0000 UTC m=+602.610026466" lastFinishedPulling="2025-11-25 10:46:48.738347893 +0000 UTC m=+605.551964490" observedRunningTime="2025-11-25 10:46:49.72978643 +0000 UTC m=+606.543403037" watchObservedRunningTime="2025-11-25 10:46:49.738044148 +0000 UTC m=+606.551660765" Nov 25 10:46:49 crc kubenswrapper[4696]: I1125 10:46:49.752166 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qt985" podStartSLOduration=2.615484758 podStartE2EDuration="5.752146025s" podCreationTimestamp="2025-11-25 10:46:44 +0000 UTC" firstStartedPulling="2025-11-25 10:46:45.53860504 +0000 UTC m=+602.352221637" lastFinishedPulling="2025-11-25 10:46:48.675266297 +0000 UTC m=+605.488882904" observedRunningTime="2025-11-25 10:46:49.747637219 +0000 UTC m=+606.561253806" watchObservedRunningTime="2025-11-25 10:46:49.752146025 +0000 UTC m=+606.565762622" Nov 25 10:46:50 crc kubenswrapper[4696]: I1125 10:46:50.723892 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" event={"ID":"ed104e46-ddf7-496c-a389-1c78d24e8e7c","Type":"ContainerStarted","Data":"94de047b8fb6c8457b7d5f21ebff5a5e148d2e369185e1c447eb7487cbd1f21c"} Nov 25 10:46:50 crc kubenswrapper[4696]: I1125 10:46:50.742899 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" podStartSLOduration=2.822891294 podStartE2EDuration="6.742883871s" podCreationTimestamp="2025-11-25 10:46:44 +0000 UTC" firstStartedPulling="2025-11-25 10:46:45.794474366 +0000 UTC m=+602.608090953" lastFinishedPulling="2025-11-25 10:46:49.714466943 +0000 UTC m=+606.528083530" observedRunningTime="2025-11-25 10:46:50.742638402 +0000 UTC m=+607.556254999" watchObservedRunningTime="2025-11-25 10:46:50.742883871 +0000 UTC m=+607.556500458" Nov 25 10:46:51 crc kubenswrapper[4696]: I1125 10:46:51.730021 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.310615 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-knqmh" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.552302 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-l8nb9"] Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.552691 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.552732 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="nbdb" containerID="cri-o://703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.552816 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="northd" containerID="cri-o://80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.553049 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-node" containerID="cri-o://cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.553188 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-acl-logging" containerID="cri-o://798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.553198 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="sbdb" containerID="cri-o://c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.553287 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-controller" containerID="cri-o://714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.591212 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" containerID="cri-o://85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" gracePeriod=30 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.754809 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/2.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.755244 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/1.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.755273 4696 generic.go:334] "Generic (PLEG): container finished" podID="bcb29cb9-408c-48c2-bdca-2c3dce47b991" containerID="0e32d100ef8aa5d1698b8603daba8da5a1f983cd8740afe907b4552254b7ee52" exitCode=2 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.755315 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerDied","Data":"0e32d100ef8aa5d1698b8603daba8da5a1f983cd8740afe907b4552254b7ee52"} Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.755371 4696 scope.go:117] "RemoveContainer" containerID="227c1bd94c66f00702ecf4f39b7912b6454731b7631b872662c8d05547250964" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.755833 4696 scope.go:117] "RemoveContainer" containerID="0e32d100ef8aa5d1698b8603daba8da5a1f983cd8740afe907b4552254b7ee52" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.756065 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-g77vb_openshift-multus(bcb29cb9-408c-48c2-bdca-2c3dce47b991)\"" pod="openshift-multus/multus-g77vb" podUID="bcb29cb9-408c-48c2-bdca-2c3dce47b991" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.759606 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/3.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.761424 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovn-acl-logging/0.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762088 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovn-controller/0.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762501 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" exitCode=0 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762523 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" exitCode=0 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762532 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" exitCode=0 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762540 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" exitCode=143 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762548 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" exitCode=143 Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762566 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef"} Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762592 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b"} Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762601 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2"} Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762610 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e"} Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.762618 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4"} Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.829410 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovnkube-controller/3.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.831254 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovn-acl-logging/0.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.831576 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovn-controller/0.log" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.831903 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.832471 4696 scope.go:117] "RemoveContainer" containerID="57ed1b7625fb7f870ee8be1ad1527d2b8e8369b37972ac0e3407d28064e117e4" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882416 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9mtw2"] Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882640 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-acl-logging" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882653 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-acl-logging" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882683 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882693 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882703 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882712 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882720 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882727 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882740 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882749 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882758 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882766 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882776 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882782 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882791 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="nbdb" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882799 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="nbdb" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882811 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-node" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882817 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-node" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882828 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="sbdb" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882837 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="sbdb" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882848 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="northd" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882855 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="northd" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882868 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882874 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: E1125 10:46:55.882885 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kubecfg-setup" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.882893 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kubecfg-setup" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883023 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883034 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883042 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-acl-logging" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883053 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883059 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="sbdb" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883071 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883083 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883095 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="northd" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883102 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="kube-rbac-proxy-node" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883110 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="nbdb" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883120 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovn-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.883309 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerName="ovnkube-controller" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.885910 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974358 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovn-node-metrics-cert\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974425 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-openvswitch\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974469 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-ovn-kubernetes\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974505 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkcm8\" (UniqueName: \"kubernetes.io/projected/f2a63620-dc3c-467f-9cf5-d688f0e233c3-kube-api-access-wkcm8\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974538 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-kubelet\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974585 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-config\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974616 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974656 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-bin\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974724 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-systemd-units\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974756 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-etc-openvswitch\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974801 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974810 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-systemd\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974842 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974852 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-log-socket\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974912 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-netd\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974870 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-log-socket" (OuterVolumeSpecName: "log-socket") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.974966 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975014 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975020 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975052 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975099 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-env-overrides\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975117 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975181 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-slash\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975241 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-ovn\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975280 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-slash" (OuterVolumeSpecName: "host-slash") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975313 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-node-log\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975334 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975362 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-netns\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975422 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-node-log" (OuterVolumeSpecName: "node-log") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975387 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-script-lib\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975463 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975478 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975471 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-var-lib-openvswitch\") pod \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\" (UID: \"f2a63620-dc3c-467f-9cf5-d688f0e233c3\") " Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975519 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975529 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975744 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975825 4696 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975844 4696 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975855 4696 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975863 4696 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975871 4696 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975879 4696 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975886 4696 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975894 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975903 4696 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975911 4696 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975919 4696 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975928 4696 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975937 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975945 4696 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975953 4696 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975962 4696 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.975970 4696 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.981546 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.982041 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2a63620-dc3c-467f-9cf5-d688f0e233c3-kube-api-access-wkcm8" (OuterVolumeSpecName: "kube-api-access-wkcm8") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "kube-api-access-wkcm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:46:55 crc kubenswrapper[4696]: I1125 10:46:55.992676 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f2a63620-dc3c-467f-9cf5-d688f0e233c3" (UID: "f2a63620-dc3c-467f-9cf5-d688f0e233c3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076499 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-var-lib-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076550 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-kubelet\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076574 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-cni-bin\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076619 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-log-socket\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076643 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-cni-netd\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076708 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovnkube-script-lib\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076734 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076811 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgk2z\" (UniqueName: \"kubernetes.io/projected/fbbffa2b-141e-462a-86c8-83be7b5a0386-kube-api-access-jgk2z\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076878 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovnkube-config\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076916 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovn-node-metrics-cert\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076953 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-ovn\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.076985 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-etc-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077022 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-node-log\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077062 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-run-ovn-kubernetes\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077102 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-systemd-units\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077137 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-run-netns\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077172 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-systemd\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077225 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-slash\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077258 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-env-overrides\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077293 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077376 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f2a63620-dc3c-467f-9cf5-d688f0e233c3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077402 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkcm8\" (UniqueName: \"kubernetes.io/projected/f2a63620-dc3c-467f-9cf5-d688f0e233c3-kube-api-access-wkcm8\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.077421 4696 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f2a63620-dc3c-467f-9cf5-d688f0e233c3-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.178825 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.178889 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-var-lib-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.178914 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-kubelet\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.178933 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-log-socket\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179091 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-cni-bin\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179002 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-var-lib-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179191 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-cni-netd\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179084 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-kubelet\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179120 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-cni-bin\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179125 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-cni-netd\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179313 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179353 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovnkube-script-lib\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179374 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgk2z\" (UniqueName: \"kubernetes.io/projected/fbbffa2b-141e-462a-86c8-83be7b5a0386-kube-api-access-jgk2z\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179410 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovnkube-config\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179435 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovn-node-metrics-cert\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179455 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-ovn\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179475 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-etc-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179497 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-node-log\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179520 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-run-ovn-kubernetes\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179629 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-ovn\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179945 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.179048 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-log-socket\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180095 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-etc-openvswitch\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180271 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-systemd-units\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180398 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-run-netns\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180584 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-systemd\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180710 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-run-netns\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180445 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-node-log\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180680 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-systemd-units\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180402 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovnkube-config\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180426 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-run-ovn-kubernetes\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.180873 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovnkube-script-lib\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.181174 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-run-systemd\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.181297 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-slash\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.181693 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-env-overrides\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.181649 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fbbffa2b-141e-462a-86c8-83be7b5a0386-host-slash\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.182090 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fbbffa2b-141e-462a-86c8-83be7b5a0386-env-overrides\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.185990 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbbffa2b-141e-462a-86c8-83be7b5a0386-ovn-node-metrics-cert\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.196127 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgk2z\" (UniqueName: \"kubernetes.io/projected/fbbffa2b-141e-462a-86c8-83be7b5a0386-kube-api-access-jgk2z\") pod \"ovnkube-node-9mtw2\" (UID: \"fbbffa2b-141e-462a-86c8-83be7b5a0386\") " pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.199217 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:46:56 crc kubenswrapper[4696]: W1125 10:46:56.230090 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbffa2b_141e_462a_86c8_83be7b5a0386.slice/crio-a9285c132546648073926f9c67e6c048a7643a411925e19e016119fa37d46c00 WatchSource:0}: Error finding container a9285c132546648073926f9c67e6c048a7643a411925e19e016119fa37d46c00: Status 404 returned error can't find the container with id a9285c132546648073926f9c67e6c048a7643a411925e19e016119fa37d46c00 Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.773611 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovn-acl-logging/0.log" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775098 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-l8nb9_f2a63620-dc3c-467f-9cf5-d688f0e233c3/ovn-controller/0.log" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775579 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" exitCode=0 Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775606 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" exitCode=0 Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775617 4696 generic.go:334] "Generic (PLEG): container finished" podID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" containerID="80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" exitCode=0 Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775731 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775721 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886"} Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775907 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd"} Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775929 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8"} Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775942 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8nb9" event={"ID":"f2a63620-dc3c-467f-9cf5-d688f0e233c3","Type":"ContainerDied","Data":"8a44975297d61e26c42357b632f58b16bebc8697ada169443d0253a9e0c95334"} Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.775962 4696 scope.go:117] "RemoveContainer" containerID="85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.777441 4696 generic.go:334] "Generic (PLEG): container finished" podID="fbbffa2b-141e-462a-86c8-83be7b5a0386" containerID="fd9aeb37836cee242c3e568c1ebf9d76f1de64769f263ef265f77ce7615011a7" exitCode=0 Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.777480 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerDied","Data":"fd9aeb37836cee242c3e568c1ebf9d76f1de64769f263ef265f77ce7615011a7"} Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.777515 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"a9285c132546648073926f9c67e6c048a7643a411925e19e016119fa37d46c00"} Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.780379 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/2.log" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.818516 4696 scope.go:117] "RemoveContainer" containerID="c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.852526 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-l8nb9"] Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.863495 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-l8nb9"] Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.869934 4696 scope.go:117] "RemoveContainer" containerID="703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.888519 4696 scope.go:117] "RemoveContainer" containerID="80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.904460 4696 scope.go:117] "RemoveContainer" containerID="09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.916004 4696 scope.go:117] "RemoveContainer" containerID="cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.928657 4696 scope.go:117] "RemoveContainer" containerID="798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.940461 4696 scope.go:117] "RemoveContainer" containerID="714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.963057 4696 scope.go:117] "RemoveContainer" containerID="89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.991713 4696 scope.go:117] "RemoveContainer" containerID="85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.992265 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": container with ID starting with 85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef not found: ID does not exist" containerID="85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.992322 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef"} err="failed to get container status \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": rpc error: code = NotFound desc = could not find container \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": container with ID starting with 85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.992356 4696 scope.go:117] "RemoveContainer" containerID="c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.993031 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": container with ID starting with c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886 not found: ID does not exist" containerID="c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.993848 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886"} err="failed to get container status \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": rpc error: code = NotFound desc = could not find container \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": container with ID starting with c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.993913 4696 scope.go:117] "RemoveContainer" containerID="703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.994442 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": container with ID starting with 703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd not found: ID does not exist" containerID="703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.994487 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd"} err="failed to get container status \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": rpc error: code = NotFound desc = could not find container \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": container with ID starting with 703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.994514 4696 scope.go:117] "RemoveContainer" containerID="80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.994913 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": container with ID starting with 80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8 not found: ID does not exist" containerID="80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.994944 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8"} err="failed to get container status \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": rpc error: code = NotFound desc = could not find container \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": container with ID starting with 80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.994967 4696 scope.go:117] "RemoveContainer" containerID="09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.995547 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": container with ID starting with 09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b not found: ID does not exist" containerID="09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.995572 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b"} err="failed to get container status \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": rpc error: code = NotFound desc = could not find container \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": container with ID starting with 09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.995593 4696 scope.go:117] "RemoveContainer" containerID="cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.995919 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": container with ID starting with cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2 not found: ID does not exist" containerID="cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.995946 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2"} err="failed to get container status \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": rpc error: code = NotFound desc = could not find container \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": container with ID starting with cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.995964 4696 scope.go:117] "RemoveContainer" containerID="798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.996367 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": container with ID starting with 798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e not found: ID does not exist" containerID="798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.996400 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e"} err="failed to get container status \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": rpc error: code = NotFound desc = could not find container \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": container with ID starting with 798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.996418 4696 scope.go:117] "RemoveContainer" containerID="714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.996714 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": container with ID starting with 714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4 not found: ID does not exist" containerID="714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.996739 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4"} err="failed to get container status \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": rpc error: code = NotFound desc = could not find container \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": container with ID starting with 714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.996752 4696 scope.go:117] "RemoveContainer" containerID="89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8" Nov 25 10:46:56 crc kubenswrapper[4696]: E1125 10:46:56.997160 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": container with ID starting with 89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8 not found: ID does not exist" containerID="89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.997191 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8"} err="failed to get container status \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": rpc error: code = NotFound desc = could not find container \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": container with ID starting with 89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.997208 4696 scope.go:117] "RemoveContainer" containerID="85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.997423 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef"} err="failed to get container status \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": rpc error: code = NotFound desc = could not find container \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": container with ID starting with 85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.997473 4696 scope.go:117] "RemoveContainer" containerID="c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.998398 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886"} err="failed to get container status \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": rpc error: code = NotFound desc = could not find container \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": container with ID starting with c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.998425 4696 scope.go:117] "RemoveContainer" containerID="703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.998768 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd"} err="failed to get container status \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": rpc error: code = NotFound desc = could not find container \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": container with ID starting with 703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.998794 4696 scope.go:117] "RemoveContainer" containerID="80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.999231 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8"} err="failed to get container status \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": rpc error: code = NotFound desc = could not find container \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": container with ID starting with 80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.999253 4696 scope.go:117] "RemoveContainer" containerID="09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.999573 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b"} err="failed to get container status \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": rpc error: code = NotFound desc = could not find container \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": container with ID starting with 09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.999615 4696 scope.go:117] "RemoveContainer" containerID="cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.999925 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2"} err="failed to get container status \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": rpc error: code = NotFound desc = could not find container \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": container with ID starting with cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2 not found: ID does not exist" Nov 25 10:46:56 crc kubenswrapper[4696]: I1125 10:46:56.999948 4696 scope.go:117] "RemoveContainer" containerID="798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.000198 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e"} err="failed to get container status \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": rpc error: code = NotFound desc = could not find container \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": container with ID starting with 798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.000228 4696 scope.go:117] "RemoveContainer" containerID="714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.000479 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4"} err="failed to get container status \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": rpc error: code = NotFound desc = could not find container \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": container with ID starting with 714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.000502 4696 scope.go:117] "RemoveContainer" containerID="89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.000882 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8"} err="failed to get container status \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": rpc error: code = NotFound desc = could not find container \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": container with ID starting with 89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.000907 4696 scope.go:117] "RemoveContainer" containerID="85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.001247 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef"} err="failed to get container status \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": rpc error: code = NotFound desc = could not find container \"85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef\": container with ID starting with 85765a0e621c3aa24bcdeebfbee55b67600ee91e3b648d3f73c2a4588a7ceeef not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.001280 4696 scope.go:117] "RemoveContainer" containerID="c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.001741 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886"} err="failed to get container status \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": rpc error: code = NotFound desc = could not find container \"c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886\": container with ID starting with c3a026bee77cd9a1c27012976dcc7d3d8c3a876bc78ffdc6f0301e22fd781886 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.001773 4696 scope.go:117] "RemoveContainer" containerID="703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.002038 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd"} err="failed to get container status \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": rpc error: code = NotFound desc = could not find container \"703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd\": container with ID starting with 703addc9dc08a85b7b242090b329f1b1c98648880dbdf6a898f0dab7aaf019cd not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.002063 4696 scope.go:117] "RemoveContainer" containerID="80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.002283 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8"} err="failed to get container status \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": rpc error: code = NotFound desc = could not find container \"80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8\": container with ID starting with 80cf42e232771f3933d438f7d4ab55c0354d565db23d73fa5b31a4dd435b0bb8 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.002315 4696 scope.go:117] "RemoveContainer" containerID="09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.002575 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b"} err="failed to get container status \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": rpc error: code = NotFound desc = could not find container \"09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b\": container with ID starting with 09d5b03f3d4b3a992f985a00bc07d0890c6144f028dfa5611ec1b4ac9207859b not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.002603 4696 scope.go:117] "RemoveContainer" containerID="cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003063 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2"} err="failed to get container status \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": rpc error: code = NotFound desc = could not find container \"cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2\": container with ID starting with cfdfda323f35f371e8ebdfc8a7aaae0ec1417cc44a72085f695502cb8fef99f2 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003093 4696 scope.go:117] "RemoveContainer" containerID="798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003342 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e"} err="failed to get container status \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": rpc error: code = NotFound desc = could not find container \"798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e\": container with ID starting with 798272581e56173a7aa1625201a70eef2d0fa6f8f562aa5b03458666fd374d1e not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003371 4696 scope.go:117] "RemoveContainer" containerID="714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003653 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4"} err="failed to get container status \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": rpc error: code = NotFound desc = could not find container \"714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4\": container with ID starting with 714ff0caa9fb1a1ab3bed797aafc5c8bd9cd95647a97a0335abfa81b7fec4ce4 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003689 4696 scope.go:117] "RemoveContainer" containerID="89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.003896 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8"} err="failed to get container status \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": rpc error: code = NotFound desc = could not find container \"89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8\": container with ID starting with 89cf94bcb59754a57a214ef4a3c5391321117b6a602049b2d5bdc93197af5de8 not found: ID does not exist" Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.789925 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"9e14fbc59cc808415e0f4ffa71f5b4d95133ee47cd02a95167056f8d31f96191"} Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.791189 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"1dcd094948c9aba1a7f601954637b94d3fec74e52ae7a32104c31643a60d3503"} Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.791271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"b92077612d586e5e489fd7aaea4470d3ac0a3d62703d464a9d7b675062798086"} Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.791356 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"b30fd748874b0ec0ca41e0563b574e235e3a17a1ee797ed2a7e03bb78cdfc24b"} Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.791426 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"e8ae38ce8ffbde135bc60ec066bb57f170eeb7761d42cf4548115afb7940dfd9"} Nov 25 10:46:57 crc kubenswrapper[4696]: I1125 10:46:57.791498 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"5418d33977c753b7edff8093651daecab4c63804f44070383d3426a5a1ac6589"} Nov 25 10:46:58 crc kubenswrapper[4696]: I1125 10:46:58.051369 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2a63620-dc3c-467f-9cf5-d688f0e233c3" path="/var/lib/kubelet/pods/f2a63620-dc3c-467f-9cf5-d688f0e233c3/volumes" Nov 25 10:47:00 crc kubenswrapper[4696]: I1125 10:47:00.802356 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:47:00 crc kubenswrapper[4696]: I1125 10:47:00.802738 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:47:00 crc kubenswrapper[4696]: I1125 10:47:00.819371 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"ab43bc6dc2a48eb0bd142807bbfda86afb92141e6fee816f5b9bad3748f6bd7c"} Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.833057 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" event={"ID":"fbbffa2b-141e-462a-86c8-83be7b5a0386","Type":"ContainerStarted","Data":"8ccb1685d28f0b30dc2e3284531ac54a95e85eaa146422f35a91b71c02028391"} Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.835629 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.835710 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.835790 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.864546 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.881328 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:47:02 crc kubenswrapper[4696]: I1125 10:47:02.896759 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" podStartSLOduration=7.896746483 podStartE2EDuration="7.896746483s" podCreationTimestamp="2025-11-25 10:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:02.872930441 +0000 UTC m=+619.686547048" watchObservedRunningTime="2025-11-25 10:47:02.896746483 +0000 UTC m=+619.710363070" Nov 25 10:47:08 crc kubenswrapper[4696]: I1125 10:47:08.983161 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-clh68"] Nov 25 10:47:08 crc kubenswrapper[4696]: I1125 10:47:08.983742 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" podUID="0be09268-4279-43e9-8da0-78c4cdd920cd" containerName="controller-manager" containerID="cri-o://e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7" gracePeriod=30 Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.108018 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn"] Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.108287 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" podUID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" containerName="route-controller-manager" containerID="cri-o://ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec" gracePeriod=30 Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.251088 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.299914 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452376 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzntr\" (UniqueName: \"kubernetes.io/projected/0be09268-4279-43e9-8da0-78c4cdd920cd-kube-api-access-mzntr\") pod \"0be09268-4279-43e9-8da0-78c4cdd920cd\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452751 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-client-ca\") pod \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452813 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-config\") pod \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452846 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-proxy-ca-bundles\") pod \"0be09268-4279-43e9-8da0-78c4cdd920cd\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452894 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89c24\" (UniqueName: \"kubernetes.io/projected/735300e4-0a1b-4e3f-9735-84abcf63cf4a-kube-api-access-89c24\") pod \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452926 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-client-ca\") pod \"0be09268-4279-43e9-8da0-78c4cdd920cd\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.452953 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735300e4-0a1b-4e3f-9735-84abcf63cf4a-serving-cert\") pod \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\" (UID: \"735300e4-0a1b-4e3f-9735-84abcf63cf4a\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.453003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-config\") pod \"0be09268-4279-43e9-8da0-78c4cdd920cd\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.453054 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be09268-4279-43e9-8da0-78c4cdd920cd-serving-cert\") pod \"0be09268-4279-43e9-8da0-78c4cdd920cd\" (UID: \"0be09268-4279-43e9-8da0-78c4cdd920cd\") " Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.454249 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-client-ca" (OuterVolumeSpecName: "client-ca") pod "735300e4-0a1b-4e3f-9735-84abcf63cf4a" (UID: "735300e4-0a1b-4e3f-9735-84abcf63cf4a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.454285 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-client-ca" (OuterVolumeSpecName: "client-ca") pod "0be09268-4279-43e9-8da0-78c4cdd920cd" (UID: "0be09268-4279-43e9-8da0-78c4cdd920cd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.454319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0be09268-4279-43e9-8da0-78c4cdd920cd" (UID: "0be09268-4279-43e9-8da0-78c4cdd920cd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.454732 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-config" (OuterVolumeSpecName: "config") pod "735300e4-0a1b-4e3f-9735-84abcf63cf4a" (UID: "735300e4-0a1b-4e3f-9735-84abcf63cf4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.455074 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-config" (OuterVolumeSpecName: "config") pod "0be09268-4279-43e9-8da0-78c4cdd920cd" (UID: "0be09268-4279-43e9-8da0-78c4cdd920cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.459119 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be09268-4279-43e9-8da0-78c4cdd920cd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0be09268-4279-43e9-8da0-78c4cdd920cd" (UID: "0be09268-4279-43e9-8da0-78c4cdd920cd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.459170 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/735300e4-0a1b-4e3f-9735-84abcf63cf4a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "735300e4-0a1b-4e3f-9735-84abcf63cf4a" (UID: "735300e4-0a1b-4e3f-9735-84abcf63cf4a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.459211 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/735300e4-0a1b-4e3f-9735-84abcf63cf4a-kube-api-access-89c24" (OuterVolumeSpecName: "kube-api-access-89c24") pod "735300e4-0a1b-4e3f-9735-84abcf63cf4a" (UID: "735300e4-0a1b-4e3f-9735-84abcf63cf4a"). InnerVolumeSpecName "kube-api-access-89c24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.459243 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be09268-4279-43e9-8da0-78c4cdd920cd-kube-api-access-mzntr" (OuterVolumeSpecName: "kube-api-access-mzntr") pod "0be09268-4279-43e9-8da0-78c4cdd920cd" (UID: "0be09268-4279-43e9-8da0-78c4cdd920cd"). InnerVolumeSpecName "kube-api-access-mzntr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554619 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554705 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be09268-4279-43e9-8da0-78c4cdd920cd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554727 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzntr\" (UniqueName: \"kubernetes.io/projected/0be09268-4279-43e9-8da0-78c4cdd920cd-kube-api-access-mzntr\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554746 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554766 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735300e4-0a1b-4e3f-9735-84abcf63cf4a-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554782 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554799 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89c24\" (UniqueName: \"kubernetes.io/projected/735300e4-0a1b-4e3f-9735-84abcf63cf4a-kube-api-access-89c24\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554815 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be09268-4279-43e9-8da0-78c4cdd920cd-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.554831 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735300e4-0a1b-4e3f-9735-84abcf63cf4a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.874778 4696 generic.go:334] "Generic (PLEG): container finished" podID="0be09268-4279-43e9-8da0-78c4cdd920cd" containerID="e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7" exitCode=0 Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.874859 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" event={"ID":"0be09268-4279-43e9-8da0-78c4cdd920cd","Type":"ContainerDied","Data":"e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7"} Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.874890 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" event={"ID":"0be09268-4279-43e9-8da0-78c4cdd920cd","Type":"ContainerDied","Data":"faff271f3ba07160ea5a28d9a5338c4c34b1d70b15548cce624a6508945648a0"} Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.874911 4696 scope.go:117] "RemoveContainer" containerID="e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.875015 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-clh68" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.877391 4696 generic.go:334] "Generic (PLEG): container finished" podID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" containerID="ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec" exitCode=0 Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.877423 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.877538 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" event={"ID":"735300e4-0a1b-4e3f-9735-84abcf63cf4a","Type":"ContainerDied","Data":"ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec"} Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.877589 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn" event={"ID":"735300e4-0a1b-4e3f-9735-84abcf63cf4a","Type":"ContainerDied","Data":"eac984141d5e6b5c91c00b360cba8b01b3588d9dc6c0b5c405ab3b7d62cf058d"} Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.924768 4696 scope.go:117] "RemoveContainer" containerID="e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7" Nov 25 10:47:09 crc kubenswrapper[4696]: E1125 10:47:09.925549 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7\": container with ID starting with e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7 not found: ID does not exist" containerID="e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.925777 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7"} err="failed to get container status \"e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7\": rpc error: code = NotFound desc = could not find container \"e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7\": container with ID starting with e1df041debdb2321a814a8cf11b644da213d0d51585cba710c48ef9c05e87bb7 not found: ID does not exist" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.926928 4696 scope.go:117] "RemoveContainer" containerID="ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.939813 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-clh68"] Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.946993 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-clh68"] Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.952762 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn"] Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.954164 4696 scope.go:117] "RemoveContainer" containerID="ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec" Nov 25 10:47:09 crc kubenswrapper[4696]: E1125 10:47:09.954691 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec\": container with ID starting with ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec not found: ID does not exist" containerID="ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.954729 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec"} err="failed to get container status \"ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec\": rpc error: code = NotFound desc = could not find container \"ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec\": container with ID starting with ed32aab86b53ff70fcd2fa49a84f846b493fb42af1e4e9f90397f3403cc1afec not found: ID does not exist" Nov 25 10:47:09 crc kubenswrapper[4696]: I1125 10:47:09.958140 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4dhfn"] Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.043361 4696 scope.go:117] "RemoveContainer" containerID="0e32d100ef8aa5d1698b8603daba8da5a1f983cd8740afe907b4552254b7ee52" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.043652 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-g77vb_openshift-multus(bcb29cb9-408c-48c2-bdca-2c3dce47b991)\"" pod="openshift-multus/multus-g77vb" podUID="bcb29cb9-408c-48c2-bdca-2c3dce47b991" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.052234 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be09268-4279-43e9-8da0-78c4cdd920cd" path="/var/lib/kubelet/pods/0be09268-4279-43e9-8da0-78c4cdd920cd/volumes" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.053041 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" path="/var/lib/kubelet/pods/735300e4-0a1b-4e3f-9735-84abcf63cf4a/volumes" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.381538 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8"] Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.381968 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" containerName="route-controller-manager" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.382005 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" containerName="route-controller-manager" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.382039 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be09268-4279-43e9-8da0-78c4cdd920cd" containerName="controller-manager" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.382056 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be09268-4279-43e9-8da0-78c4cdd920cd" containerName="controller-manager" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.382249 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be09268-4279-43e9-8da0-78c4cdd920cd" containerName="controller-manager" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.382285 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="735300e4-0a1b-4e3f-9735-84abcf63cf4a" containerName="route-controller-manager" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.382909 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.388578 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.388597 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.388619 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.388587 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.389805 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.390714 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.400065 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.400288 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8"] Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.408968 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw"] Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.410488 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.422812 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.423049 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.423190 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.423299 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.423311 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.423223 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.439998 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw"] Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465422 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3734bb4a-a107-4fa1-aa15-23bcfe179854-serving-cert\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d44920f6-aa3b-48d2-835f-5491ddb1bd57-serving-cert\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-config\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465652 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-client-ca\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465726 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-config\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465747 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw9g7\" (UniqueName: \"kubernetes.io/projected/d44920f6-aa3b-48d2-835f-5491ddb1bd57-kube-api-access-jw9g7\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465805 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-client-ca\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.465856 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-proxy-ca-bundles\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566472 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3734bb4a-a107-4fa1-aa15-23bcfe179854-serving-cert\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566547 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d44920f6-aa3b-48d2-835f-5491ddb1bd57-serving-cert\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566571 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-config\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566601 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnwht\" (UniqueName: \"kubernetes.io/projected/3734bb4a-a107-4fa1-aa15-23bcfe179854-kube-api-access-jnwht\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566624 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-client-ca\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566652 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-config\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw9g7\" (UniqueName: \"kubernetes.io/projected/d44920f6-aa3b-48d2-835f-5491ddb1bd57-kube-api-access-jw9g7\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566797 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-client-ca\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.566858 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-proxy-ca-bundles\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.567542 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-client-ca\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.567998 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-client-ca\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.568004 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-proxy-ca-bundles\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.568104 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-config\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.569031 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-config\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.570654 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d44920f6-aa3b-48d2-835f-5491ddb1bd57-serving-cert\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.571709 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3734bb4a-a107-4fa1-aa15-23bcfe179854-serving-cert\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.587953 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw9g7\" (UniqueName: \"kubernetes.io/projected/d44920f6-aa3b-48d2-835f-5491ddb1bd57-kube-api-access-jw9g7\") pod \"route-controller-manager-555d48bddd-wxzgw\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.667583 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnwht\" (UniqueName: \"kubernetes.io/projected/3734bb4a-a107-4fa1-aa15-23bcfe179854-kube-api-access-jnwht\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.685107 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnwht\" (UniqueName: \"kubernetes.io/projected/3734bb4a-a107-4fa1-aa15-23bcfe179854-kube-api-access-jnwht\") pod \"controller-manager-6b85fbb4c8-rsqb8\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.698572 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.726135 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1bd64530b07bd4af43b83e8d4934eb0aa5b7e5908f4e59feb612f81c0bca7b5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.726421 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1bd64530b07bd4af43b83e8d4934eb0aa5b7e5908f4e59feb612f81c0bca7b5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.726448 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1bd64530b07bd4af43b83e8d4934eb0aa5b7e5908f4e59feb612f81c0bca7b5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.726507 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager(3734bb4a-a107-4fa1-aa15-23bcfe179854)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager(3734bb4a-a107-4fa1-aa15-23bcfe179854)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1bd64530b07bd4af43b83e8d4934eb0aa5b7e5908f4e59feb612f81c0bca7b5f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.736474 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.758731 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(247d5c45ab36ff2907365ef3de2829d8ee3c911d7a016f846d131f96d180cc75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.758899 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(247d5c45ab36ff2907365ef3de2829d8ee3c911d7a016f846d131f96d180cc75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.759004 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(247d5c45ab36ff2907365ef3de2829d8ee3c911d7a016f846d131f96d180cc75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.759134 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager(d44920f6-aa3b-48d2-835f-5491ddb1bd57)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager(d44920f6-aa3b-48d2-835f-5491ddb1bd57)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(247d5c45ab36ff2907365ef3de2829d8ee3c911d7a016f846d131f96d180cc75): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.887845 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.888157 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.888232 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: I1125 10:47:10.889355 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.933801 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1fd1ccff734adacc55b65c151f8bf8c7626bf1ec4896d13800e316c5ab302373): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.933954 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1fd1ccff734adacc55b65c151f8bf8c7626bf1ec4896d13800e316c5ab302373): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.934036 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1fd1ccff734adacc55b65c151f8bf8c7626bf1ec4896d13800e316c5ab302373): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.934139 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager(3734bb4a-a107-4fa1-aa15-23bcfe179854)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager(3734bb4a-a107-4fa1-aa15-23bcfe179854)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(1fd1ccff734adacc55b65c151f8bf8c7626bf1ec4896d13800e316c5ab302373): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.940644 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(1a6c9e0bc855e51bd81c575d9ae052e4f9a0aff15fd5a7fef4cce3c90c1a0b08): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.940819 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(1a6c9e0bc855e51bd81c575d9ae052e4f9a0aff15fd5a7fef4cce3c90c1a0b08): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.940878 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(1a6c9e0bc855e51bd81c575d9ae052e4f9a0aff15fd5a7fef4cce3c90c1a0b08): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:10 crc kubenswrapper[4696]: E1125 10:47:10.940937 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager(d44920f6-aa3b-48d2-835f-5491ddb1bd57)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager(d44920f6-aa3b-48d2-835f-5491ddb1bd57)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-555d48bddd-wxzgw_openshift-route-controller-manager_d44920f6-aa3b-48d2-835f-5491ddb1bd57_0(1a6c9e0bc855e51bd81c575d9ae052e4f9a0aff15fd5a7fef4cce3c90c1a0b08): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" Nov 25 10:47:23 crc kubenswrapper[4696]: I1125 10:47:23.041887 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:23 crc kubenswrapper[4696]: I1125 10:47:23.042691 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:23 crc kubenswrapper[4696]: E1125 10:47:23.079733 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(88ee7420aeff9ef768c49f1bad38b34d2c37cf8cf11bfc9326ea1d1e765d365f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:47:23 crc kubenswrapper[4696]: E1125 10:47:23.079804 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(88ee7420aeff9ef768c49f1bad38b34d2c37cf8cf11bfc9326ea1d1e765d365f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:23 crc kubenswrapper[4696]: E1125 10:47:23.079832 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(88ee7420aeff9ef768c49f1bad38b34d2c37cf8cf11bfc9326ea1d1e765d365f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:23 crc kubenswrapper[4696]: E1125 10:47:23.079894 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager(3734bb4a-a107-4fa1-aa15-23bcfe179854)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager(3734bb4a-a107-4fa1-aa15-23bcfe179854)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6b85fbb4c8-rsqb8_openshift-controller-manager_3734bb4a-a107-4fa1-aa15-23bcfe179854_0(88ee7420aeff9ef768c49f1bad38b34d2c37cf8cf11bfc9326ea1d1e765d365f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" Nov 25 10:47:24 crc kubenswrapper[4696]: I1125 10:47:24.046941 4696 scope.go:117] "RemoveContainer" containerID="0e32d100ef8aa5d1698b8603daba8da5a1f983cd8740afe907b4552254b7ee52" Nov 25 10:47:24 crc kubenswrapper[4696]: I1125 10:47:24.970510 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g77vb_bcb29cb9-408c-48c2-bdca-2c3dce47b991/kube-multus/2.log" Nov 25 10:47:24 crc kubenswrapper[4696]: I1125 10:47:24.970961 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g77vb" event={"ID":"bcb29cb9-408c-48c2-bdca-2c3dce47b991","Type":"ContainerStarted","Data":"f52d8bc8dd2fb1eb66dd0afbad0fec8d733d110cdfc7e2f04d3d3b6047c49542"} Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.041976 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.042753 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.229379 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9mtw2" Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.299726 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw"] Nov 25 10:47:26 crc kubenswrapper[4696]: W1125 10:47:26.308725 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd44920f6_aa3b_48d2_835f_5491ddb1bd57.slice/crio-71491ebdea5a79ff83767f40e2c9b663f758524b2b9740a73f344d910a5a7b6f WatchSource:0}: Error finding container 71491ebdea5a79ff83767f40e2c9b663f758524b2b9740a73f344d910a5a7b6f: Status 404 returned error can't find the container with id 71491ebdea5a79ff83767f40e2c9b663f758524b2b9740a73f344d910a5a7b6f Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.994037 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" event={"ID":"d44920f6-aa3b-48d2-835f-5491ddb1bd57","Type":"ContainerStarted","Data":"2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42"} Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.994384 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" event={"ID":"d44920f6-aa3b-48d2-835f-5491ddb1bd57","Type":"ContainerStarted","Data":"71491ebdea5a79ff83767f40e2c9b663f758524b2b9740a73f344d910a5a7b6f"} Nov 25 10:47:26 crc kubenswrapper[4696]: I1125 10:47:26.994808 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:27 crc kubenswrapper[4696]: I1125 10:47:27.019013 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" podStartSLOduration=17.018991505 podStartE2EDuration="17.018991505s" podCreationTimestamp="2025-11-25 10:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:27.018032333 +0000 UTC m=+643.831648920" watchObservedRunningTime="2025-11-25 10:47:27.018991505 +0000 UTC m=+643.832608092" Nov 25 10:47:27 crc kubenswrapper[4696]: I1125 10:47:27.149170 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:30 crc kubenswrapper[4696]: I1125 10:47:30.802063 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:47:30 crc kubenswrapper[4696]: I1125 10:47:30.802500 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:47:30 crc kubenswrapper[4696]: I1125 10:47:30.802553 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:47:30 crc kubenswrapper[4696]: I1125 10:47:30.803182 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e613ae8723bcb59d28226fb02f66886a567bae947befcc699ab9dd24e0648204"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:47:30 crc kubenswrapper[4696]: I1125 10:47:30.803241 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://e613ae8723bcb59d28226fb02f66886a567bae947befcc699ab9dd24e0648204" gracePeriod=600 Nov 25 10:47:32 crc kubenswrapper[4696]: I1125 10:47:32.024807 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="e613ae8723bcb59d28226fb02f66886a567bae947befcc699ab9dd24e0648204" exitCode=0 Nov 25 10:47:32 crc kubenswrapper[4696]: I1125 10:47:32.024861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"e613ae8723bcb59d28226fb02f66886a567bae947befcc699ab9dd24e0648204"} Nov 25 10:47:32 crc kubenswrapper[4696]: I1125 10:47:32.025504 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"e7253a2ef071b30ac37f2b8fc9846ceb3e4a68877fb90b28472f1abfa37caf36"} Nov 25 10:47:32 crc kubenswrapper[4696]: I1125 10:47:32.025529 4696 scope.go:117] "RemoveContainer" containerID="1fb36885e86a8194f5415b7d52233333dec593850e5cf8d383f363bb7a6ad19c" Nov 25 10:47:34 crc kubenswrapper[4696]: I1125 10:47:34.045396 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:34 crc kubenswrapper[4696]: I1125 10:47:34.045973 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:34 crc kubenswrapper[4696]: I1125 10:47:34.459016 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8"] Nov 25 10:47:35 crc kubenswrapper[4696]: I1125 10:47:35.045227 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" event={"ID":"3734bb4a-a107-4fa1-aa15-23bcfe179854","Type":"ContainerStarted","Data":"df161de579d34af7e69602e68b7dc48b851530331a960e3c647abbeb8c56ec6f"} Nov 25 10:47:35 crc kubenswrapper[4696]: I1125 10:47:35.045536 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" event={"ID":"3734bb4a-a107-4fa1-aa15-23bcfe179854","Type":"ContainerStarted","Data":"aa5b96f42f0f8f2b5bd1e9582fcb3bf1cf4e79cc203f78ae8bf8477b37d92179"} Nov 25 10:47:35 crc kubenswrapper[4696]: I1125 10:47:35.046501 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:35 crc kubenswrapper[4696]: I1125 10:47:35.054847 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:47:35 crc kubenswrapper[4696]: I1125 10:47:35.073505 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" podStartSLOduration=25.073485515 podStartE2EDuration="25.073485515s" podCreationTimestamp="2025-11-25 10:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:35.070468829 +0000 UTC m=+651.884085416" watchObservedRunningTime="2025-11-25 10:47:35.073485515 +0000 UTC m=+651.887102102" Nov 25 10:47:48 crc kubenswrapper[4696]: I1125 10:47:48.998047 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw"] Nov 25 10:47:48 crc kubenswrapper[4696]: I1125 10:47:48.999711 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" containerName="route-controller-manager" containerID="cri-o://2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42" gracePeriod=30 Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.437090 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.638580 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-config\") pod \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.638658 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw9g7\" (UniqueName: \"kubernetes.io/projected/d44920f6-aa3b-48d2-835f-5491ddb1bd57-kube-api-access-jw9g7\") pod \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.638748 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d44920f6-aa3b-48d2-835f-5491ddb1bd57-serving-cert\") pod \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.638785 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-client-ca\") pod \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\" (UID: \"d44920f6-aa3b-48d2-835f-5491ddb1bd57\") " Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.639467 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-client-ca" (OuterVolumeSpecName: "client-ca") pod "d44920f6-aa3b-48d2-835f-5491ddb1bd57" (UID: "d44920f6-aa3b-48d2-835f-5491ddb1bd57"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.639778 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-config" (OuterVolumeSpecName: "config") pod "d44920f6-aa3b-48d2-835f-5491ddb1bd57" (UID: "d44920f6-aa3b-48d2-835f-5491ddb1bd57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.646766 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d44920f6-aa3b-48d2-835f-5491ddb1bd57-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d44920f6-aa3b-48d2-835f-5491ddb1bd57" (UID: "d44920f6-aa3b-48d2-835f-5491ddb1bd57"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.647254 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d44920f6-aa3b-48d2-835f-5491ddb1bd57-kube-api-access-jw9g7" (OuterVolumeSpecName: "kube-api-access-jw9g7") pod "d44920f6-aa3b-48d2-835f-5491ddb1bd57" (UID: "d44920f6-aa3b-48d2-835f-5491ddb1bd57"). InnerVolumeSpecName "kube-api-access-jw9g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.740649 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.740744 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d44920f6-aa3b-48d2-835f-5491ddb1bd57-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.740772 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw9g7\" (UniqueName: \"kubernetes.io/projected/d44920f6-aa3b-48d2-835f-5491ddb1bd57-kube-api-access-jw9g7\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:49 crc kubenswrapper[4696]: I1125 10:47:49.740798 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d44920f6-aa3b-48d2-835f-5491ddb1bd57-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.134421 4696 generic.go:334] "Generic (PLEG): container finished" podID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" containerID="2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42" exitCode=0 Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.134488 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" event={"ID":"d44920f6-aa3b-48d2-835f-5491ddb1bd57","Type":"ContainerDied","Data":"2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42"} Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.134527 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" event={"ID":"d44920f6-aa3b-48d2-835f-5491ddb1bd57","Type":"ContainerDied","Data":"71491ebdea5a79ff83767f40e2c9b663f758524b2b9740a73f344d910a5a7b6f"} Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.134556 4696 scope.go:117] "RemoveContainer" containerID="2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.134798 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.160162 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw"] Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.164509 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-555d48bddd-wxzgw"] Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.165703 4696 scope.go:117] "RemoveContainer" containerID="2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42" Nov 25 10:47:50 crc kubenswrapper[4696]: E1125 10:47:50.166450 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42\": container with ID starting with 2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42 not found: ID does not exist" containerID="2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.166519 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42"} err="failed to get container status \"2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42\": rpc error: code = NotFound desc = could not find container \"2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42\": container with ID starting with 2a07a0680c37c895dc40a1d081feadde3ad134823086ab1ea1719f101a409c42 not found: ID does not exist" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.748952 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k"] Nov 25 10:47:50 crc kubenswrapper[4696]: E1125 10:47:50.749289 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" containerName="route-controller-manager" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.749308 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" containerName="route-controller-manager" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.749492 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" containerName="route-controller-manager" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.750064 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.753333 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.753580 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.753871 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.754101 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.754383 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.755355 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.767599 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k"] Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.856231 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-config\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.856287 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-serving-cert\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.856309 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-client-ca\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.856330 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhrf9\" (UniqueName: \"kubernetes.io/projected/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-kube-api-access-jhrf9\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.957540 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-config\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.957655 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-serving-cert\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.957742 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-client-ca\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.957797 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhrf9\" (UniqueName: \"kubernetes.io/projected/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-kube-api-access-jhrf9\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.959138 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-config\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.959464 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-client-ca\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.961878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-serving-cert\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:50 crc kubenswrapper[4696]: I1125 10:47:50.983998 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhrf9\" (UniqueName: \"kubernetes.io/projected/f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5-kube-api-access-jhrf9\") pod \"route-controller-manager-7f9d5d569b-wkd6k\" (UID: \"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5\") " pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:51 crc kubenswrapper[4696]: I1125 10:47:51.066651 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:51 crc kubenswrapper[4696]: I1125 10:47:51.476355 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k"] Nov 25 10:47:52 crc kubenswrapper[4696]: I1125 10:47:52.051338 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d44920f6-aa3b-48d2-835f-5491ddb1bd57" path="/var/lib/kubelet/pods/d44920f6-aa3b-48d2-835f-5491ddb1bd57/volumes" Nov 25 10:47:52 crc kubenswrapper[4696]: I1125 10:47:52.147490 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" event={"ID":"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5","Type":"ContainerStarted","Data":"fa00761f44111fe50dff5c862943040606e0ffb7ab667c36be3f36ac766d638a"} Nov 25 10:47:52 crc kubenswrapper[4696]: I1125 10:47:52.147842 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" event={"ID":"f9df0f19-7e9e-4e2c-bf7f-8c2712d9e4f5","Type":"ContainerStarted","Data":"813220b85a9ab867967a62271ac8b9dd5c269bf1623407672e3acb99ebeaeac0"} Nov 25 10:47:52 crc kubenswrapper[4696]: I1125 10:47:52.148394 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:52 crc kubenswrapper[4696]: I1125 10:47:52.153232 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" Nov 25 10:47:52 crc kubenswrapper[4696]: I1125 10:47:52.166770 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7f9d5d569b-wkd6k" podStartSLOduration=3.166755346 podStartE2EDuration="3.166755346s" podCreationTimestamp="2025-11-25 10:47:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:52.165548808 +0000 UTC m=+668.979165405" watchObservedRunningTime="2025-11-25 10:47:52.166755346 +0000 UTC m=+668.980371933" Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.733760 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn"] Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.736862 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.742208 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn"] Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.743761 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.920825 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qh92\" (UniqueName: \"kubernetes.io/projected/7cce94e7-26a7-4e28-8880-7a47b2d5116f-kube-api-access-7qh92\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.920989 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:03 crc kubenswrapper[4696]: I1125 10:48:03.921045 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.022617 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.022747 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.022823 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qh92\" (UniqueName: \"kubernetes.io/projected/7cce94e7-26a7-4e28-8880-7a47b2d5116f-kube-api-access-7qh92\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.023408 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.023837 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.061429 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qh92\" (UniqueName: \"kubernetes.io/projected/7cce94e7-26a7-4e28-8880-7a47b2d5116f-kube-api-access-7qh92\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.061744 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:04 crc kubenswrapper[4696]: I1125 10:48:04.482894 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn"] Nov 25 10:48:05 crc kubenswrapper[4696]: I1125 10:48:05.219622 4696 generic.go:334] "Generic (PLEG): container finished" podID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerID="7a061824e164bfadd2bfb3c178ebb3f983433053b48f4574fa5d37ff4f2dd167" exitCode=0 Nov 25 10:48:05 crc kubenswrapper[4696]: I1125 10:48:05.219725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" event={"ID":"7cce94e7-26a7-4e28-8880-7a47b2d5116f","Type":"ContainerDied","Data":"7a061824e164bfadd2bfb3c178ebb3f983433053b48f4574fa5d37ff4f2dd167"} Nov 25 10:48:05 crc kubenswrapper[4696]: I1125 10:48:05.219942 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" event={"ID":"7cce94e7-26a7-4e28-8880-7a47b2d5116f","Type":"ContainerStarted","Data":"1cf07f8eeb8dfd2d4a8dab341bf68e6ddc01ffee6bd112e3857d9f94ff7ac8c4"} Nov 25 10:48:07 crc kubenswrapper[4696]: I1125 10:48:07.234918 4696 generic.go:334] "Generic (PLEG): container finished" podID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerID="00d0563ca4c49f43e48d6d8d549f357fe1e85ebcc296fd0f432f81700005350a" exitCode=0 Nov 25 10:48:07 crc kubenswrapper[4696]: I1125 10:48:07.234990 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" event={"ID":"7cce94e7-26a7-4e28-8880-7a47b2d5116f","Type":"ContainerDied","Data":"00d0563ca4c49f43e48d6d8d549f357fe1e85ebcc296fd0f432f81700005350a"} Nov 25 10:48:08 crc kubenswrapper[4696]: I1125 10:48:08.247355 4696 generic.go:334] "Generic (PLEG): container finished" podID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerID="7804ba8009a4a682ff0b0269422e3040e2c8226bbfe5bb3fbaeff19638ccc30b" exitCode=0 Nov 25 10:48:08 crc kubenswrapper[4696]: I1125 10:48:08.247435 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" event={"ID":"7cce94e7-26a7-4e28-8880-7a47b2d5116f","Type":"ContainerDied","Data":"7804ba8009a4a682ff0b0269422e3040e2c8226bbfe5bb3fbaeff19638ccc30b"} Nov 25 10:48:08 crc kubenswrapper[4696]: I1125 10:48:08.954556 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8"] Nov 25 10:48:08 crc kubenswrapper[4696]: I1125 10:48:08.954831 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" containerName="controller-manager" containerID="cri-o://df161de579d34af7e69602e68b7dc48b851530331a960e3c647abbeb8c56ec6f" gracePeriod=30 Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.257494 4696 generic.go:334] "Generic (PLEG): container finished" podID="3734bb4a-a107-4fa1-aa15-23bcfe179854" containerID="df161de579d34af7e69602e68b7dc48b851530331a960e3c647abbeb8c56ec6f" exitCode=0 Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.257545 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" event={"ID":"3734bb4a-a107-4fa1-aa15-23bcfe179854","Type":"ContainerDied","Data":"df161de579d34af7e69602e68b7dc48b851530331a960e3c647abbeb8c56ec6f"} Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.403727 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.474895 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.600710 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-config\") pod \"3734bb4a-a107-4fa1-aa15-23bcfe179854\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.600822 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qh92\" (UniqueName: \"kubernetes.io/projected/7cce94e7-26a7-4e28-8880-7a47b2d5116f-kube-api-access-7qh92\") pod \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.601829 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-client-ca\") pod \"3734bb4a-a107-4fa1-aa15-23bcfe179854\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.601902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-proxy-ca-bundles\") pod \"3734bb4a-a107-4fa1-aa15-23bcfe179854\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602021 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3734bb4a-a107-4fa1-aa15-23bcfe179854-serving-cert\") pod \"3734bb4a-a107-4fa1-aa15-23bcfe179854\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.601970 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-config" (OuterVolumeSpecName: "config") pod "3734bb4a-a107-4fa1-aa15-23bcfe179854" (UID: "3734bb4a-a107-4fa1-aa15-23bcfe179854"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602079 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-util\") pod \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602105 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-bundle\") pod \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\" (UID: \"7cce94e7-26a7-4e28-8880-7a47b2d5116f\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602152 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnwht\" (UniqueName: \"kubernetes.io/projected/3734bb4a-a107-4fa1-aa15-23bcfe179854-kube-api-access-jnwht\") pod \"3734bb4a-a107-4fa1-aa15-23bcfe179854\" (UID: \"3734bb4a-a107-4fa1-aa15-23bcfe179854\") " Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-client-ca" (OuterVolumeSpecName: "client-ca") pod "3734bb4a-a107-4fa1-aa15-23bcfe179854" (UID: "3734bb4a-a107-4fa1-aa15-23bcfe179854"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602551 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3734bb4a-a107-4fa1-aa15-23bcfe179854" (UID: "3734bb4a-a107-4fa1-aa15-23bcfe179854"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.602879 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-bundle" (OuterVolumeSpecName: "bundle") pod "7cce94e7-26a7-4e28-8880-7a47b2d5116f" (UID: "7cce94e7-26a7-4e28-8880-7a47b2d5116f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.603204 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.603216 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.603225 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.603233 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3734bb4a-a107-4fa1-aa15-23bcfe179854-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.606791 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3734bb4a-a107-4fa1-aa15-23bcfe179854-kube-api-access-jnwht" (OuterVolumeSpecName: "kube-api-access-jnwht") pod "3734bb4a-a107-4fa1-aa15-23bcfe179854" (UID: "3734bb4a-a107-4fa1-aa15-23bcfe179854"). InnerVolumeSpecName "kube-api-access-jnwht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.606841 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cce94e7-26a7-4e28-8880-7a47b2d5116f-kube-api-access-7qh92" (OuterVolumeSpecName: "kube-api-access-7qh92") pod "7cce94e7-26a7-4e28-8880-7a47b2d5116f" (UID: "7cce94e7-26a7-4e28-8880-7a47b2d5116f"). InnerVolumeSpecName "kube-api-access-7qh92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.607751 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3734bb4a-a107-4fa1-aa15-23bcfe179854-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3734bb4a-a107-4fa1-aa15-23bcfe179854" (UID: "3734bb4a-a107-4fa1-aa15-23bcfe179854"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.630514 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-util" (OuterVolumeSpecName: "util") pod "7cce94e7-26a7-4e28-8880-7a47b2d5116f" (UID: "7cce94e7-26a7-4e28-8880-7a47b2d5116f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.704813 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qh92\" (UniqueName: \"kubernetes.io/projected/7cce94e7-26a7-4e28-8880-7a47b2d5116f-kube-api-access-7qh92\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.704860 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3734bb4a-a107-4fa1-aa15-23bcfe179854-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.704881 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7cce94e7-26a7-4e28-8880-7a47b2d5116f-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4696]: I1125 10:48:09.704904 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnwht\" (UniqueName: \"kubernetes.io/projected/3734bb4a-a107-4fa1-aa15-23bcfe179854-kube-api-access-jnwht\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.264942 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" event={"ID":"3734bb4a-a107-4fa1-aa15-23bcfe179854","Type":"ContainerDied","Data":"aa5b96f42f0f8f2b5bd1e9582fcb3bf1cf4e79cc203f78ae8bf8477b37d92179"} Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.265125 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.265285 4696 scope.go:117] "RemoveContainer" containerID="df161de579d34af7e69602e68b7dc48b851530331a960e3c647abbeb8c56ec6f" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.270008 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.270195 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn" event={"ID":"7cce94e7-26a7-4e28-8880-7a47b2d5116f","Type":"ContainerDied","Data":"1cf07f8eeb8dfd2d4a8dab341bf68e6ddc01ffee6bd112e3857d9f94ff7ac8c4"} Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.270783 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf07f8eeb8dfd2d4a8dab341bf68e6ddc01ffee6bd112e3857d9f94ff7ac8c4" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.301306 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8"] Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.301368 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b85fbb4c8-rsqb8"] Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750028 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd7b978-62jr2"] Nov 25 10:48:10 crc kubenswrapper[4696]: E1125 10:48:10.750375 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" containerName="controller-manager" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750404 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" containerName="controller-manager" Nov 25 10:48:10 crc kubenswrapper[4696]: E1125 10:48:10.750422 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="pull" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750434 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="pull" Nov 25 10:48:10 crc kubenswrapper[4696]: E1125 10:48:10.750458 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="util" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750470 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="util" Nov 25 10:48:10 crc kubenswrapper[4696]: E1125 10:48:10.750488 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="extract" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750500 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="extract" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750691 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" containerName="controller-manager" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.750716 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cce94e7-26a7-4e28-8880-7a47b2d5116f" containerName="extract" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.751314 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.754455 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.754770 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.755108 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.755424 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.756008 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.756151 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.764565 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd7b978-62jr2"] Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.773190 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.818458 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-client-ca\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.818526 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9xld\" (UniqueName: \"kubernetes.io/projected/234af183-8c2d-4522-ac22-658b4c4f3fb1-kube-api-access-s9xld\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.818569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-proxy-ca-bundles\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.818606 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-config\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.818629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/234af183-8c2d-4522-ac22-658b4c4f3fb1-serving-cert\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.919717 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-config\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.919766 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/234af183-8c2d-4522-ac22-658b4c4f3fb1-serving-cert\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.919836 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-client-ca\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.919867 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9xld\" (UniqueName: \"kubernetes.io/projected/234af183-8c2d-4522-ac22-658b4c4f3fb1-kube-api-access-s9xld\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.919900 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-proxy-ca-bundles\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.921453 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-proxy-ca-bundles\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.922122 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-client-ca\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.922920 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/234af183-8c2d-4522-ac22-658b4c4f3fb1-config\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.924010 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/234af183-8c2d-4522-ac22-658b4c4f3fb1-serving-cert\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:10 crc kubenswrapper[4696]: I1125 10:48:10.938205 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9xld\" (UniqueName: \"kubernetes.io/projected/234af183-8c2d-4522-ac22-658b4c4f3fb1-kube-api-access-s9xld\") pod \"controller-manager-6c5bd7b978-62jr2\" (UID: \"234af183-8c2d-4522-ac22-658b4c4f3fb1\") " pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:11 crc kubenswrapper[4696]: I1125 10:48:11.071515 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:11 crc kubenswrapper[4696]: I1125 10:48:11.287805 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd7b978-62jr2"] Nov 25 10:48:12 crc kubenswrapper[4696]: I1125 10:48:12.050437 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3734bb4a-a107-4fa1-aa15-23bcfe179854" path="/var/lib/kubelet/pods/3734bb4a-a107-4fa1-aa15-23bcfe179854/volumes" Nov 25 10:48:12 crc kubenswrapper[4696]: I1125 10:48:12.287861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" event={"ID":"234af183-8c2d-4522-ac22-658b4c4f3fb1","Type":"ContainerStarted","Data":"436e9176e8d7d8ec50f9b6edcc45c85cd3f906e2ce5e929f3622edd21b0d8e5d"} Nov 25 10:48:12 crc kubenswrapper[4696]: I1125 10:48:12.287943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" event={"ID":"234af183-8c2d-4522-ac22-658b4c4f3fb1","Type":"ContainerStarted","Data":"aaaf1f667f81ce11afbb8ac4020eb8fbb922f3fcfa3e27cdf5737a1e5e17b46c"} Nov 25 10:48:12 crc kubenswrapper[4696]: I1125 10:48:12.288416 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:12 crc kubenswrapper[4696]: I1125 10:48:12.298248 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" Nov 25 10:48:12 crc kubenswrapper[4696]: I1125 10:48:12.308272 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6c5bd7b978-62jr2" podStartSLOduration=4.308218762 podStartE2EDuration="4.308218762s" podCreationTimestamp="2025-11-25 10:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:12.306444055 +0000 UTC m=+689.120060682" watchObservedRunningTime="2025-11-25 10:48:12.308218762 +0000 UTC m=+689.121835389" Nov 25 10:48:15 crc kubenswrapper[4696]: I1125 10:48:15.924565 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-wkpl2"] Nov 25 10:48:15 crc kubenswrapper[4696]: I1125 10:48:15.925492 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" Nov 25 10:48:15 crc kubenswrapper[4696]: I1125 10:48:15.929750 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-qf8ns" Nov 25 10:48:15 crc kubenswrapper[4696]: I1125 10:48:15.929921 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 10:48:15 crc kubenswrapper[4696]: I1125 10:48:15.930821 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 10:48:15 crc kubenswrapper[4696]: I1125 10:48:15.949036 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-wkpl2"] Nov 25 10:48:16 crc kubenswrapper[4696]: I1125 10:48:16.081502 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh9wt\" (UniqueName: \"kubernetes.io/projected/d246ee4c-dd53-42a6-b24f-5e34687b6c8e-kube-api-access-mh9wt\") pod \"nmstate-operator-557fdffb88-wkpl2\" (UID: \"d246ee4c-dd53-42a6-b24f-5e34687b6c8e\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" Nov 25 10:48:16 crc kubenswrapper[4696]: I1125 10:48:16.182357 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh9wt\" (UniqueName: \"kubernetes.io/projected/d246ee4c-dd53-42a6-b24f-5e34687b6c8e-kube-api-access-mh9wt\") pod \"nmstate-operator-557fdffb88-wkpl2\" (UID: \"d246ee4c-dd53-42a6-b24f-5e34687b6c8e\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" Nov 25 10:48:16 crc kubenswrapper[4696]: I1125 10:48:16.199593 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh9wt\" (UniqueName: \"kubernetes.io/projected/d246ee4c-dd53-42a6-b24f-5e34687b6c8e-kube-api-access-mh9wt\") pod \"nmstate-operator-557fdffb88-wkpl2\" (UID: \"d246ee4c-dd53-42a6-b24f-5e34687b6c8e\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" Nov 25 10:48:16 crc kubenswrapper[4696]: I1125 10:48:16.239911 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" Nov 25 10:48:16 crc kubenswrapper[4696]: I1125 10:48:16.657065 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-wkpl2"] Nov 25 10:48:16 crc kubenswrapper[4696]: W1125 10:48:16.661289 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd246ee4c_dd53_42a6_b24f_5e34687b6c8e.slice/crio-68007a0f58b15611ad2eaa3084d94a213bd6b2c59592fac0925f2f21ade84ba0 WatchSource:0}: Error finding container 68007a0f58b15611ad2eaa3084d94a213bd6b2c59592fac0925f2f21ade84ba0: Status 404 returned error can't find the container with id 68007a0f58b15611ad2eaa3084d94a213bd6b2c59592fac0925f2f21ade84ba0 Nov 25 10:48:17 crc kubenswrapper[4696]: I1125 10:48:17.321330 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" event={"ID":"d246ee4c-dd53-42a6-b24f-5e34687b6c8e","Type":"ContainerStarted","Data":"68007a0f58b15611ad2eaa3084d94a213bd6b2c59592fac0925f2f21ade84ba0"} Nov 25 10:48:22 crc kubenswrapper[4696]: I1125 10:48:22.363538 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" event={"ID":"d246ee4c-dd53-42a6-b24f-5e34687b6c8e","Type":"ContainerStarted","Data":"4854950a0a5e6bca67450eb628f5f1f9b432f071749614d30dab22acc536b15a"} Nov 25 10:48:22 crc kubenswrapper[4696]: I1125 10:48:22.388492 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-wkpl2" podStartSLOduration=2.953321937 podStartE2EDuration="7.388467975s" podCreationTimestamp="2025-11-25 10:48:15 +0000 UTC" firstStartedPulling="2025-11-25 10:48:16.663231417 +0000 UTC m=+693.476848004" lastFinishedPulling="2025-11-25 10:48:21.098377435 +0000 UTC m=+697.911994042" observedRunningTime="2025-11-25 10:48:22.38455532 +0000 UTC m=+699.198171927" watchObservedRunningTime="2025-11-25 10:48:22.388467975 +0000 UTC m=+699.202084572" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.268426 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.269837 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.273604 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-lxzd2" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.292575 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-ttnxf"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.293292 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.299362 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.300469 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.302304 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.354700 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.376746 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421249 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whdv7\" (UniqueName: \"kubernetes.io/projected/585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d-kube-api-access-whdv7\") pod \"nmstate-metrics-5dcf9c57c5-xznfr\" (UID: \"585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421695 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwtjl\" (UniqueName: \"kubernetes.io/projected/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-kube-api-access-vwtjl\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421798 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-dbus-socket\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421862 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-ovs-socket\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421895 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55v9k\" (UniqueName: \"kubernetes.io/projected/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-kube-api-access-55v9k\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421925 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-nmstate-lock\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.421953 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.435458 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.436113 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.438987 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.439080 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.439301 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-lk7k5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.449651 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523012 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwtjl\" (UniqueName: \"kubernetes.io/projected/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-kube-api-access-vwtjl\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523169 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-dbus-socket\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523224 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-ovs-socket\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523254 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55v9k\" (UniqueName: \"kubernetes.io/projected/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-kube-api-access-55v9k\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523280 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-nmstate-lock\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523309 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523343 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whdv7\" (UniqueName: \"kubernetes.io/projected/585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d-kube-api-access-whdv7\") pod \"nmstate-metrics-5dcf9c57c5-xznfr\" (UID: \"585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523471 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-ovs-socket\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.523915 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-nmstate-lock\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: E1125 10:48:27.524017 4696 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 25 10:48:27 crc kubenswrapper[4696]: E1125 10:48:27.524088 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-tls-key-pair podName:fc74e56a-3f49-4d38-b5e4-78cafb3276a6 nodeName:}" failed. No retries permitted until 2025-11-25 10:48:28.024065675 +0000 UTC m=+704.837682282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-tls-key-pair") pod "nmstate-webhook-6b89b748d8-5bpk8" (UID: "fc74e56a-3f49-4d38-b5e4-78cafb3276a6") : secret "openshift-nmstate-webhook" not found Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.524091 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-dbus-socket\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.546280 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwtjl\" (UniqueName: \"kubernetes.io/projected/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-kube-api-access-vwtjl\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.549799 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55v9k\" (UniqueName: \"kubernetes.io/projected/a2a61ba7-423c-4fbc-9bf5-63af41ee1170-kube-api-access-55v9k\") pod \"nmstate-handler-ttnxf\" (UID: \"a2a61ba7-423c-4fbc-9bf5-63af41ee1170\") " pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.550327 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whdv7\" (UniqueName: \"kubernetes.io/projected/585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d-kube-api-access-whdv7\") pod \"nmstate-metrics-5dcf9c57c5-xznfr\" (UID: \"585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.589183 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.622781 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.628314 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf7l5\" (UniqueName: \"kubernetes.io/projected/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-kube-api-access-lf7l5\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.628397 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.628420 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.681067 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-54bcc64569-srv6b"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.681761 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.699077 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54bcc64569-srv6b"] Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.729728 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-serving-cert\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.729769 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.729790 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-config\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.729814 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730618 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730613 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcw9b\" (UniqueName: \"kubernetes.io/projected/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-kube-api-access-kcw9b\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730673 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-oauth-config\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-service-ca\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730771 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-oauth-serving-cert\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730855 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf7l5\" (UniqueName: \"kubernetes.io/projected/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-kube-api-access-lf7l5\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.730893 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-trusted-ca-bundle\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.734388 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.757253 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf7l5\" (UniqueName: \"kubernetes.io/projected/b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332-kube-api-access-lf7l5\") pod \"nmstate-console-plugin-5874bd7bc5-52cq5\" (UID: \"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.831238 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-service-ca\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.831534 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-oauth-serving-cert\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.832121 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-service-ca\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.832372 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-oauth-serving-cert\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.832496 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-trusted-ca-bundle\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.832598 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-serving-cert\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.832686 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-config\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.832773 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcw9b\" (UniqueName: \"kubernetes.io/projected/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-kube-api-access-kcw9b\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.833588 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-oauth-config\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.833419 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-config\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.833556 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-trusted-ca-bundle\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.836525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-oauth-config\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.837620 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-console-serving-cert\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.852113 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcw9b\" (UniqueName: \"kubernetes.io/projected/4cb24163-a5d7-4b5c-be1d-c0a71d664bee-kube-api-access-kcw9b\") pod \"console-54bcc64569-srv6b\" (UID: \"4cb24163-a5d7-4b5c-be1d-c0a71d664bee\") " pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:27 crc kubenswrapper[4696]: I1125 10:48:27.995560 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.036837 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.039986 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc74e56a-3f49-4d38-b5e4-78cafb3276a6-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-5bpk8\" (UID: \"fc74e56a-3f49-4d38-b5e4-78cafb3276a6\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.052046 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.081391 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr"] Nov 25 10:48:28 crc kubenswrapper[4696]: W1125 10:48:28.100284 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod585d3d73_0434_4ce9_b2e7_04bdaf2e7c1d.slice/crio-0a018bdade9b301bf99bcd3c1d60cfba7e209965cd0461d4db35f51e62d80b4f WatchSource:0}: Error finding container 0a018bdade9b301bf99bcd3c1d60cfba7e209965cd0461d4db35f51e62d80b4f: Status 404 returned error can't find the container with id 0a018bdade9b301bf99bcd3c1d60cfba7e209965cd0461d4db35f51e62d80b4f Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.231531 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.401718 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ttnxf" event={"ID":"a2a61ba7-423c-4fbc-9bf5-63af41ee1170","Type":"ContainerStarted","Data":"6b3c2b5d60fad5e426e4727ce30600a7b9c43d489ef5d4837f62b11079f65fdf"} Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.404333 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" event={"ID":"585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d","Type":"ContainerStarted","Data":"0a018bdade9b301bf99bcd3c1d60cfba7e209965cd0461d4db35f51e62d80b4f"} Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.419454 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54bcc64569-srv6b"] Nov 25 10:48:28 crc kubenswrapper[4696]: W1125 10:48:28.424332 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cb24163_a5d7_4b5c_be1d_c0a71d664bee.slice/crio-38af50cdebdc807fc7ba7c9210187acdfde52ef8ad44c8d4c70fbb85e941eabd WatchSource:0}: Error finding container 38af50cdebdc807fc7ba7c9210187acdfde52ef8ad44c8d4c70fbb85e941eabd: Status 404 returned error can't find the container with id 38af50cdebdc807fc7ba7c9210187acdfde52ef8ad44c8d4c70fbb85e941eabd Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.504440 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5"] Nov 25 10:48:28 crc kubenswrapper[4696]: W1125 10:48:28.510905 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb35b7a4c_2b0c_47ed_8c2f_3cb5a5d11332.slice/crio-5ca130d621eb55bb1bc7bdeede3f7cb5a797fad22170194709f69c961e3d8e6e WatchSource:0}: Error finding container 5ca130d621eb55bb1bc7bdeede3f7cb5a797fad22170194709f69c961e3d8e6e: Status 404 returned error can't find the container with id 5ca130d621eb55bb1bc7bdeede3f7cb5a797fad22170194709f69c961e3d8e6e Nov 25 10:48:28 crc kubenswrapper[4696]: I1125 10:48:28.613529 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8"] Nov 25 10:48:29 crc kubenswrapper[4696]: I1125 10:48:29.414861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54bcc64569-srv6b" event={"ID":"4cb24163-a5d7-4b5c-be1d-c0a71d664bee","Type":"ContainerStarted","Data":"fc68c5bee4db7bab6f804fc4446f3553aa4ba75a5b0a85360b5951ce23f58dc8"} Nov 25 10:48:29 crc kubenswrapper[4696]: I1125 10:48:29.414905 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54bcc64569-srv6b" event={"ID":"4cb24163-a5d7-4b5c-be1d-c0a71d664bee","Type":"ContainerStarted","Data":"38af50cdebdc807fc7ba7c9210187acdfde52ef8ad44c8d4c70fbb85e941eabd"} Nov 25 10:48:29 crc kubenswrapper[4696]: I1125 10:48:29.418172 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" event={"ID":"fc74e56a-3f49-4d38-b5e4-78cafb3276a6","Type":"ContainerStarted","Data":"2a50f6498dae7caeabef07774ae69c350add089cebcdec1ddf8dd8f581b3d3e2"} Nov 25 10:48:29 crc kubenswrapper[4696]: I1125 10:48:29.419906 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" event={"ID":"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332","Type":"ContainerStarted","Data":"5ca130d621eb55bb1bc7bdeede3f7cb5a797fad22170194709f69c961e3d8e6e"} Nov 25 10:48:29 crc kubenswrapper[4696]: I1125 10:48:29.436818 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54bcc64569-srv6b" podStartSLOduration=2.436799859 podStartE2EDuration="2.436799859s" podCreationTimestamp="2025-11-25 10:48:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:29.436691485 +0000 UTC m=+706.250308142" watchObservedRunningTime="2025-11-25 10:48:29.436799859 +0000 UTC m=+706.250416446" Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.444326 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" event={"ID":"585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d","Type":"ContainerStarted","Data":"d7a7fcdbd246a141038e01d6fd91bf8529ba504ff3865fa0b096773e3799f7f1"} Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.446078 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ttnxf" event={"ID":"a2a61ba7-423c-4fbc-9bf5-63af41ee1170","Type":"ContainerStarted","Data":"f2249f2aed54334219cb317315afb4ec727e93569bca41a39f4ccc8303c4a2ce"} Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.446683 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.448587 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" event={"ID":"fc74e56a-3f49-4d38-b5e4-78cafb3276a6","Type":"ContainerStarted","Data":"3f5b81d9b084aa003c9479a1fe59588609a77d4731c15ce4a2a279ddb2ba2ab8"} Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.449171 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.464954 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-ttnxf" podStartSLOduration=1.7109143439999999 podStartE2EDuration="4.464923304s" podCreationTimestamp="2025-11-25 10:48:27 +0000 UTC" firstStartedPulling="2025-11-25 10:48:27.654228338 +0000 UTC m=+704.467844935" lastFinishedPulling="2025-11-25 10:48:30.408237318 +0000 UTC m=+707.221853895" observedRunningTime="2025-11-25 10:48:31.461431882 +0000 UTC m=+708.275048469" watchObservedRunningTime="2025-11-25 10:48:31.464923304 +0000 UTC m=+708.278539891" Nov 25 10:48:31 crc kubenswrapper[4696]: I1125 10:48:31.485863 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" podStartSLOduration=2.6980390549999997 podStartE2EDuration="4.485835833s" podCreationTimestamp="2025-11-25 10:48:27 +0000 UTC" firstStartedPulling="2025-11-25 10:48:28.623400845 +0000 UTC m=+705.437017432" lastFinishedPulling="2025-11-25 10:48:30.411197623 +0000 UTC m=+707.224814210" observedRunningTime="2025-11-25 10:48:31.483929382 +0000 UTC m=+708.297545989" watchObservedRunningTime="2025-11-25 10:48:31.485835833 +0000 UTC m=+708.299452420" Nov 25 10:48:32 crc kubenswrapper[4696]: I1125 10:48:32.456923 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" event={"ID":"b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332","Type":"ContainerStarted","Data":"42cc6aeeec624aaed19add750d59e8ba9915126d59f1e68c03a6058a1112276e"} Nov 25 10:48:34 crc kubenswrapper[4696]: I1125 10:48:34.077841 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-52cq5" podStartSLOduration=4.251683744 podStartE2EDuration="7.077825933s" podCreationTimestamp="2025-11-25 10:48:27 +0000 UTC" firstStartedPulling="2025-11-25 10:48:28.512815187 +0000 UTC m=+705.326431774" lastFinishedPulling="2025-11-25 10:48:31.338957386 +0000 UTC m=+708.152573963" observedRunningTime="2025-11-25 10:48:32.474558175 +0000 UTC m=+709.288174782" watchObservedRunningTime="2025-11-25 10:48:34.077825933 +0000 UTC m=+710.891442520" Nov 25 10:48:34 crc kubenswrapper[4696]: I1125 10:48:34.470276 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" event={"ID":"585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d","Type":"ContainerStarted","Data":"74490f372a0b17fbff45592b86f7f6568cd6bad89e753c8cb04c60d553e38518"} Nov 25 10:48:34 crc kubenswrapper[4696]: I1125 10:48:34.489772 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xznfr" podStartSLOduration=2.211800222 podStartE2EDuration="7.489751866s" podCreationTimestamp="2025-11-25 10:48:27 +0000 UTC" firstStartedPulling="2025-11-25 10:48:28.102814595 +0000 UTC m=+704.916431182" lastFinishedPulling="2025-11-25 10:48:33.380766239 +0000 UTC m=+710.194382826" observedRunningTime="2025-11-25 10:48:34.487292538 +0000 UTC m=+711.300909125" watchObservedRunningTime="2025-11-25 10:48:34.489751866 +0000 UTC m=+711.303368453" Nov 25 10:48:37 crc kubenswrapper[4696]: I1125 10:48:37.647157 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-ttnxf" Nov 25 10:48:37 crc kubenswrapper[4696]: I1125 10:48:37.996374 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:37 crc kubenswrapper[4696]: I1125 10:48:37.996901 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:38 crc kubenswrapper[4696]: I1125 10:48:38.006611 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:38 crc kubenswrapper[4696]: I1125 10:48:38.499808 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54bcc64569-srv6b" Nov 25 10:48:38 crc kubenswrapper[4696]: I1125 10:48:38.553900 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rj85t"] Nov 25 10:48:48 crc kubenswrapper[4696]: I1125 10:48:48.239498 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-5bpk8" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.124899 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72"] Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.126568 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.128884 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.143688 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72"] Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.181928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.182133 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.182184 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk4h7\" (UniqueName: \"kubernetes.io/projected/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-kube-api-access-wk4h7\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.283642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk4h7\" (UniqueName: \"kubernetes.io/projected/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-kube-api-access-wk4h7\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.283774 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.283878 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.284407 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.284520 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.302610 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk4h7\" (UniqueName: \"kubernetes.io/projected/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-kube-api-access-wk4h7\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.444076 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:02 crc kubenswrapper[4696]: I1125 10:49:02.822878 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72"] Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.602472 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-rj85t" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerName="console" containerID="cri-o://1d793054554ae20fae52561dbb6c5ee7c3a0fea6bb62172540aea264b9b1bf2b" gracePeriod=15 Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.735278 4696 generic.go:334] "Generic (PLEG): container finished" podID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerID="777115a7fea054de0afb0013ba3697c94b56961d35ab81f6d60c85c932f73068" exitCode=0 Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.735356 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" event={"ID":"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4","Type":"ContainerDied","Data":"777115a7fea054de0afb0013ba3697c94b56961d35ab81f6d60c85c932f73068"} Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.735392 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" event={"ID":"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4","Type":"ContainerStarted","Data":"b44bcb58f273c9c12dd777d43a540f5652f792ef2074d90d521f699d04a82e4a"} Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.738568 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rj85t_376ea9c4-9e68-4475-b205-95e2185e51ba/console/0.log" Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.738644 4696 generic.go:334] "Generic (PLEG): container finished" podID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerID="1d793054554ae20fae52561dbb6c5ee7c3a0fea6bb62172540aea264b9b1bf2b" exitCode=2 Nov 25 10:49:03 crc kubenswrapper[4696]: I1125 10:49:03.738732 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rj85t" event={"ID":"376ea9c4-9e68-4475-b205-95e2185e51ba","Type":"ContainerDied","Data":"1d793054554ae20fae52561dbb6c5ee7c3a0fea6bb62172540aea264b9b1bf2b"} Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.024633 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rj85t_376ea9c4-9e68-4475-b205-95e2185e51ba/console/0.log" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.025106 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104156 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-oauth-config\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104258 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-oauth-serving-cert\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104314 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-console-config\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104399 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-service-ca\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104449 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-serving-cert\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104489 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-trusted-ca-bundle\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.104520 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7vh9\" (UniqueName: \"kubernetes.io/projected/376ea9c4-9e68-4475-b205-95e2185e51ba-kube-api-access-w7vh9\") pod \"376ea9c4-9e68-4475-b205-95e2185e51ba\" (UID: \"376ea9c4-9e68-4475-b205-95e2185e51ba\") " Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.105189 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.105424 4696 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.105451 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-service-ca" (OuterVolumeSpecName: "service-ca") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.105651 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.105671 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-console-config" (OuterVolumeSpecName: "console-config") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.109858 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376ea9c4-9e68-4475-b205-95e2185e51ba-kube-api-access-w7vh9" (OuterVolumeSpecName: "kube-api-access-w7vh9") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "kube-api-access-w7vh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.109983 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.111757 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "376ea9c4-9e68-4475-b205-95e2185e51ba" (UID: "376ea9c4-9e68-4475-b205-95e2185e51ba"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.206702 4696 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.206927 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.207003 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7vh9\" (UniqueName: \"kubernetes.io/projected/376ea9c4-9e68-4475-b205-95e2185e51ba-kube-api-access-w7vh9\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.207106 4696 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/376ea9c4-9e68-4475-b205-95e2185e51ba-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.207164 4696 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.207234 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/376ea9c4-9e68-4475-b205-95e2185e51ba-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.745272 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rj85t_376ea9c4-9e68-4475-b205-95e2185e51ba/console/0.log" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.745325 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rj85t" event={"ID":"376ea9c4-9e68-4475-b205-95e2185e51ba","Type":"ContainerDied","Data":"1e2cd1359cb85fa71f7451b4377204d89df1d5c5a90787309f387ad09c232586"} Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.745364 4696 scope.go:117] "RemoveContainer" containerID="1d793054554ae20fae52561dbb6c5ee7c3a0fea6bb62172540aea264b9b1bf2b" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.745419 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rj85t" Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.777256 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rj85t"] Nov 25 10:49:04 crc kubenswrapper[4696]: I1125 10:49:04.781863 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-rj85t"] Nov 25 10:49:05 crc kubenswrapper[4696]: I1125 10:49:05.752401 4696 generic.go:334] "Generic (PLEG): container finished" podID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerID="3e9f5eb9fec5f0480e7d656d359ba1c7724d75439d142f886c25765fa7808e4a" exitCode=0 Nov 25 10:49:05 crc kubenswrapper[4696]: I1125 10:49:05.752457 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" event={"ID":"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4","Type":"ContainerDied","Data":"3e9f5eb9fec5f0480e7d656d359ba1c7724d75439d142f886c25765fa7808e4a"} Nov 25 10:49:06 crc kubenswrapper[4696]: I1125 10:49:06.049500 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" path="/var/lib/kubelet/pods/376ea9c4-9e68-4475-b205-95e2185e51ba/volumes" Nov 25 10:49:06 crc kubenswrapper[4696]: I1125 10:49:06.763082 4696 generic.go:334] "Generic (PLEG): container finished" podID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerID="b69a633f63aeee30931a0a9500a08edf0607d636e54f1e98951e7b4641a17ba4" exitCode=0 Nov 25 10:49:06 crc kubenswrapper[4696]: I1125 10:49:06.763151 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" event={"ID":"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4","Type":"ContainerDied","Data":"b69a633f63aeee30931a0a9500a08edf0607d636e54f1e98951e7b4641a17ba4"} Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.026131 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.156844 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk4h7\" (UniqueName: \"kubernetes.io/projected/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-kube-api-access-wk4h7\") pod \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.157243 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-util\") pod \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.157265 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-bundle\") pod \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\" (UID: \"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4\") " Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.158509 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-bundle" (OuterVolumeSpecName: "bundle") pod "ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" (UID: "ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.163045 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-kube-api-access-wk4h7" (OuterVolumeSpecName: "kube-api-access-wk4h7") pod "ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" (UID: "ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4"). InnerVolumeSpecName "kube-api-access-wk4h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.187986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-util" (OuterVolumeSpecName: "util") pod "ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" (UID: "ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.259270 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk4h7\" (UniqueName: \"kubernetes.io/projected/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-kube-api-access-wk4h7\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.259341 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.259357 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.777712 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" event={"ID":"ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4","Type":"ContainerDied","Data":"b44bcb58f273c9c12dd777d43a540f5652f792ef2074d90d521f699d04a82e4a"} Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.777753 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72" Nov 25 10:49:08 crc kubenswrapper[4696]: I1125 10:49:08.777764 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b44bcb58f273c9c12dd777d43a540f5652f792ef2074d90d521f699d04a82e4a" Nov 25 10:49:20 crc kubenswrapper[4696]: I1125 10:49:20.810371 4696 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.624933 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v"] Nov 25 10:49:22 crc kubenswrapper[4696]: E1125 10:49:22.625336 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="extract" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625347 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="extract" Nov 25 10:49:22 crc kubenswrapper[4696]: E1125 10:49:22.625357 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="util" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625364 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="util" Nov 25 10:49:22 crc kubenswrapper[4696]: E1125 10:49:22.625373 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerName="console" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625379 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerName="console" Nov 25 10:49:22 crc kubenswrapper[4696]: E1125 10:49:22.625389 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="pull" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625395 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="pull" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625482 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="376ea9c4-9e68-4475-b205-95e2185e51ba" containerName="console" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625501 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4" containerName="extract" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.625875 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.629318 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-sglkm" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.629521 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.629647 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.629775 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.629892 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.654491 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v"] Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.736984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-webhook-cert\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.737035 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-apiservice-cert\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.737129 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87pd2\" (UniqueName: \"kubernetes.io/projected/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-kube-api-access-87pd2\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.838202 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-webhook-cert\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.838513 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-apiservice-cert\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.838654 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87pd2\" (UniqueName: \"kubernetes.io/projected/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-kube-api-access-87pd2\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.843929 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-webhook-cert\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.845339 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-apiservice-cert\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.865893 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87pd2\" (UniqueName: \"kubernetes.io/projected/5ad9cb9e-8e80-41bc-bb0a-687f580830b0-kube-api-access-87pd2\") pod \"metallb-operator-controller-manager-d5845fdbd-84v4v\" (UID: \"5ad9cb9e-8e80-41bc-bb0a-687f580830b0\") " pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.947492 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.990606 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5"] Nov 25 10:49:22 crc kubenswrapper[4696]: I1125 10:49:22.991798 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.001261 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xkqff" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.001707 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.001953 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.021857 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5"] Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.042966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c8010a11-0748-4b8b-b40d-14d57fe53f40-apiservice-cert\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.043028 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c8010a11-0748-4b8b-b40d-14d57fe53f40-webhook-cert\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.043073 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmpwv\" (UniqueName: \"kubernetes.io/projected/c8010a11-0748-4b8b-b40d-14d57fe53f40-kube-api-access-mmpwv\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.143681 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c8010a11-0748-4b8b-b40d-14d57fe53f40-apiservice-cert\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.143730 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c8010a11-0748-4b8b-b40d-14d57fe53f40-webhook-cert\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.143762 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmpwv\" (UniqueName: \"kubernetes.io/projected/c8010a11-0748-4b8b-b40d-14d57fe53f40-kube-api-access-mmpwv\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.150364 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c8010a11-0748-4b8b-b40d-14d57fe53f40-webhook-cert\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.153366 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c8010a11-0748-4b8b-b40d-14d57fe53f40-apiservice-cert\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.178033 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmpwv\" (UniqueName: \"kubernetes.io/projected/c8010a11-0748-4b8b-b40d-14d57fe53f40-kube-api-access-mmpwv\") pod \"metallb-operator-webhook-server-5b88d4b6cb-m8sz5\" (UID: \"c8010a11-0748-4b8b-b40d-14d57fe53f40\") " pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.386308 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.567709 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v"] Nov 25 10:49:23 crc kubenswrapper[4696]: W1125 10:49:23.583316 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ad9cb9e_8e80_41bc_bb0a_687f580830b0.slice/crio-20597e371ae67cc7dd1004cf5e75ae5c52c97182997113be9c2350e2ff97883a WatchSource:0}: Error finding container 20597e371ae67cc7dd1004cf5e75ae5c52c97182997113be9c2350e2ff97883a: Status 404 returned error can't find the container with id 20597e371ae67cc7dd1004cf5e75ae5c52c97182997113be9c2350e2ff97883a Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.602048 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5"] Nov 25 10:49:23 crc kubenswrapper[4696]: W1125 10:49:23.609054 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8010a11_0748_4b8b_b40d_14d57fe53f40.slice/crio-757adee636ea45cdffa9a7d2ee5f3f4413465ccf6952f949f191325808ee2cfc WatchSource:0}: Error finding container 757adee636ea45cdffa9a7d2ee5f3f4413465ccf6952f949f191325808ee2cfc: Status 404 returned error can't find the container with id 757adee636ea45cdffa9a7d2ee5f3f4413465ccf6952f949f191325808ee2cfc Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.859766 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" event={"ID":"c8010a11-0748-4b8b-b40d-14d57fe53f40","Type":"ContainerStarted","Data":"757adee636ea45cdffa9a7d2ee5f3f4413465ccf6952f949f191325808ee2cfc"} Nov 25 10:49:23 crc kubenswrapper[4696]: I1125 10:49:23.861125 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" event={"ID":"5ad9cb9e-8e80-41bc-bb0a-687f580830b0","Type":"ContainerStarted","Data":"20597e371ae67cc7dd1004cf5e75ae5c52c97182997113be9c2350e2ff97883a"} Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.285105 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xcz6c"] Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.286422 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.293456 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xcz6c"] Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.398692 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-catalog-content\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.398799 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcs7z\" (UniqueName: \"kubernetes.io/projected/da5dae3b-c7d7-4236-bf10-7802551338dc-kube-api-access-pcs7z\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.398823 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-utilities\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.500763 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcs7z\" (UniqueName: \"kubernetes.io/projected/da5dae3b-c7d7-4236-bf10-7802551338dc-kube-api-access-pcs7z\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.501293 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-utilities\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.501424 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-utilities\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.501594 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-catalog-content\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.502160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-catalog-content\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.521095 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcs7z\" (UniqueName: \"kubernetes.io/projected/da5dae3b-c7d7-4236-bf10-7802551338dc-kube-api-access-pcs7z\") pod \"community-operators-xcz6c\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:27 crc kubenswrapper[4696]: I1125 10:49:27.604225 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:28 crc kubenswrapper[4696]: I1125 10:49:28.893056 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" event={"ID":"c8010a11-0748-4b8b-b40d-14d57fe53f40","Type":"ContainerStarted","Data":"17037a9f6931bbe3958f2b8a0b8e2e59f60a63833e7dc2b14e339131830ce5f7"} Nov 25 10:49:28 crc kubenswrapper[4696]: I1125 10:49:28.893430 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:28 crc kubenswrapper[4696]: I1125 10:49:28.894296 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" event={"ID":"5ad9cb9e-8e80-41bc-bb0a-687f580830b0","Type":"ContainerStarted","Data":"f697d0741c6aba614f0928425b09324906e9eefff45b7de1cf2f6cc03a55f8dc"} Nov 25 10:49:28 crc kubenswrapper[4696]: I1125 10:49:28.894497 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:49:28 crc kubenswrapper[4696]: I1125 10:49:28.915705 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" podStartSLOduration=1.916993684 podStartE2EDuration="6.915686904s" podCreationTimestamp="2025-11-25 10:49:22 +0000 UTC" firstStartedPulling="2025-11-25 10:49:23.618373752 +0000 UTC m=+760.431990339" lastFinishedPulling="2025-11-25 10:49:28.617066972 +0000 UTC m=+765.430683559" observedRunningTime="2025-11-25 10:49:28.912116338 +0000 UTC m=+765.725732935" watchObservedRunningTime="2025-11-25 10:49:28.915686904 +0000 UTC m=+765.729303501" Nov 25 10:49:28 crc kubenswrapper[4696]: I1125 10:49:28.938335 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" podStartSLOduration=1.957797008 podStartE2EDuration="6.938317693s" podCreationTimestamp="2025-11-25 10:49:22 +0000 UTC" firstStartedPulling="2025-11-25 10:49:23.587960542 +0000 UTC m=+760.401577129" lastFinishedPulling="2025-11-25 10:49:28.568481227 +0000 UTC m=+765.382097814" observedRunningTime="2025-11-25 10:49:28.937187346 +0000 UTC m=+765.750803953" watchObservedRunningTime="2025-11-25 10:49:28.938317693 +0000 UTC m=+765.751934280" Nov 25 10:49:29 crc kubenswrapper[4696]: I1125 10:49:29.049993 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xcz6c"] Nov 25 10:49:29 crc kubenswrapper[4696]: I1125 10:49:29.902094 4696 generic.go:334] "Generic (PLEG): container finished" podID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerID="75028753c458aeb167becbd619603127cefa98daa51e42e6a36fe4d567ee7e61" exitCode=0 Nov 25 10:49:29 crc kubenswrapper[4696]: I1125 10:49:29.902166 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xcz6c" event={"ID":"da5dae3b-c7d7-4236-bf10-7802551338dc","Type":"ContainerDied","Data":"75028753c458aeb167becbd619603127cefa98daa51e42e6a36fe4d567ee7e61"} Nov 25 10:49:29 crc kubenswrapper[4696]: I1125 10:49:29.902843 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xcz6c" event={"ID":"da5dae3b-c7d7-4236-bf10-7802551338dc","Type":"ContainerStarted","Data":"8f432da36d5a6f1ff1ce1520529994574deaf3b1185edd375134368448e18cc9"} Nov 25 10:49:30 crc kubenswrapper[4696]: I1125 10:49:30.909279 4696 generic.go:334] "Generic (PLEG): container finished" podID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerID="f66a8469d77c6d6c96cea4cee07d8676399f390e94fd032c7094560b2b524ddd" exitCode=0 Nov 25 10:49:30 crc kubenswrapper[4696]: I1125 10:49:30.909346 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xcz6c" event={"ID":"da5dae3b-c7d7-4236-bf10-7802551338dc","Type":"ContainerDied","Data":"f66a8469d77c6d6c96cea4cee07d8676399f390e94fd032c7094560b2b524ddd"} Nov 25 10:49:31 crc kubenswrapper[4696]: I1125 10:49:31.917592 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xcz6c" event={"ID":"da5dae3b-c7d7-4236-bf10-7802551338dc","Type":"ContainerStarted","Data":"fc5760a599a4ed277b9316717ddea596cefa13aa29e345f16dcaf3563a16f441"} Nov 25 10:49:31 crc kubenswrapper[4696]: I1125 10:49:31.941365 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xcz6c" podStartSLOduration=3.531969829 podStartE2EDuration="4.941345936s" podCreationTimestamp="2025-11-25 10:49:27 +0000 UTC" firstStartedPulling="2025-11-25 10:49:29.904023037 +0000 UTC m=+766.717639624" lastFinishedPulling="2025-11-25 10:49:31.313399144 +0000 UTC m=+768.127015731" observedRunningTime="2025-11-25 10:49:31.935471946 +0000 UTC m=+768.749088533" watchObservedRunningTime="2025-11-25 10:49:31.941345936 +0000 UTC m=+768.754962523" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.295294 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9v4x4"] Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.297417 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.311705 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9v4x4"] Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.401126 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-catalog-content\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.401189 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-utilities\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.401267 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdghr\" (UniqueName: \"kubernetes.io/projected/1ea618ec-6350-4ea6-8292-d11b54f87c1c-kube-api-access-hdghr\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.502975 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-catalog-content\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.503040 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-utilities\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.503126 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdghr\" (UniqueName: \"kubernetes.io/projected/1ea618ec-6350-4ea6-8292-d11b54f87c1c-kube-api-access-hdghr\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.503525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-catalog-content\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.503751 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-utilities\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.529127 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdghr\" (UniqueName: \"kubernetes.io/projected/1ea618ec-6350-4ea6-8292-d11b54f87c1c-kube-api-access-hdghr\") pod \"redhat-marketplace-9v4x4\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:35 crc kubenswrapper[4696]: I1125 10:49:35.617967 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:36 crc kubenswrapper[4696]: I1125 10:49:36.066129 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9v4x4"] Nov 25 10:49:36 crc kubenswrapper[4696]: I1125 10:49:36.949058 4696 generic.go:334] "Generic (PLEG): container finished" podID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerID="d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72" exitCode=0 Nov 25 10:49:36 crc kubenswrapper[4696]: I1125 10:49:36.949121 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9v4x4" event={"ID":"1ea618ec-6350-4ea6-8292-d11b54f87c1c","Type":"ContainerDied","Data":"d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72"} Nov 25 10:49:36 crc kubenswrapper[4696]: I1125 10:49:36.949368 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9v4x4" event={"ID":"1ea618ec-6350-4ea6-8292-d11b54f87c1c","Type":"ContainerStarted","Data":"bd7c025794ddee9ed34c0c7f90d7b7ffc9d3ffe306815ad77f191a79bbd62034"} Nov 25 10:49:37 crc kubenswrapper[4696]: I1125 10:49:37.605298 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:37 crc kubenswrapper[4696]: I1125 10:49:37.605348 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:37 crc kubenswrapper[4696]: I1125 10:49:37.655047 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:38 crc kubenswrapper[4696]: I1125 10:49:38.001103 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:39 crc kubenswrapper[4696]: I1125 10:49:39.969970 4696 generic.go:334] "Generic (PLEG): container finished" podID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerID="e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61" exitCode=0 Nov 25 10:49:39 crc kubenswrapper[4696]: I1125 10:49:39.970042 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9v4x4" event={"ID":"1ea618ec-6350-4ea6-8292-d11b54f87c1c","Type":"ContainerDied","Data":"e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61"} Nov 25 10:49:40 crc kubenswrapper[4696]: I1125 10:49:40.977240 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9v4x4" event={"ID":"1ea618ec-6350-4ea6-8292-d11b54f87c1c","Type":"ContainerStarted","Data":"e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6"} Nov 25 10:49:41 crc kubenswrapper[4696]: I1125 10:49:40.999984 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9v4x4" podStartSLOduration=2.5725201159999997 podStartE2EDuration="5.999966804s" podCreationTimestamp="2025-11-25 10:49:35 +0000 UTC" firstStartedPulling="2025-11-25 10:49:36.950287979 +0000 UTC m=+773.763904566" lastFinishedPulling="2025-11-25 10:49:40.377734677 +0000 UTC m=+777.191351254" observedRunningTime="2025-11-25 10:49:40.997817445 +0000 UTC m=+777.811434042" watchObservedRunningTime="2025-11-25 10:49:40.999966804 +0000 UTC m=+777.813583391" Nov 25 10:49:41 crc kubenswrapper[4696]: I1125 10:49:41.276761 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xcz6c"] Nov 25 10:49:41 crc kubenswrapper[4696]: I1125 10:49:41.277206 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xcz6c" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="registry-server" containerID="cri-o://fc5760a599a4ed277b9316717ddea596cefa13aa29e345f16dcaf3563a16f441" gracePeriod=2 Nov 25 10:49:41 crc kubenswrapper[4696]: I1125 10:49:41.986476 4696 generic.go:334] "Generic (PLEG): container finished" podID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerID="fc5760a599a4ed277b9316717ddea596cefa13aa29e345f16dcaf3563a16f441" exitCode=0 Nov 25 10:49:41 crc kubenswrapper[4696]: I1125 10:49:41.986563 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xcz6c" event={"ID":"da5dae3b-c7d7-4236-bf10-7802551338dc","Type":"ContainerDied","Data":"fc5760a599a4ed277b9316717ddea596cefa13aa29e345f16dcaf3563a16f441"} Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.131168 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.187272 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcs7z\" (UniqueName: \"kubernetes.io/projected/da5dae3b-c7d7-4236-bf10-7802551338dc-kube-api-access-pcs7z\") pod \"da5dae3b-c7d7-4236-bf10-7802551338dc\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.187328 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-utilities\") pod \"da5dae3b-c7d7-4236-bf10-7802551338dc\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.187415 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-catalog-content\") pod \"da5dae3b-c7d7-4236-bf10-7802551338dc\" (UID: \"da5dae3b-c7d7-4236-bf10-7802551338dc\") " Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.188297 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-utilities" (OuterVolumeSpecName: "utilities") pod "da5dae3b-c7d7-4236-bf10-7802551338dc" (UID: "da5dae3b-c7d7-4236-bf10-7802551338dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.192726 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da5dae3b-c7d7-4236-bf10-7802551338dc-kube-api-access-pcs7z" (OuterVolumeSpecName: "kube-api-access-pcs7z") pod "da5dae3b-c7d7-4236-bf10-7802551338dc" (UID: "da5dae3b-c7d7-4236-bf10-7802551338dc"). InnerVolumeSpecName "kube-api-access-pcs7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.236199 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da5dae3b-c7d7-4236-bf10-7802551338dc" (UID: "da5dae3b-c7d7-4236-bf10-7802551338dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.289345 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcs7z\" (UniqueName: \"kubernetes.io/projected/da5dae3b-c7d7-4236-bf10-7802551338dc-kube-api-access-pcs7z\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.289390 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.289403 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5dae3b-c7d7-4236-bf10-7802551338dc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.995232 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xcz6c" event={"ID":"da5dae3b-c7d7-4236-bf10-7802551338dc","Type":"ContainerDied","Data":"8f432da36d5a6f1ff1ce1520529994574deaf3b1185edd375134368448e18cc9"} Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.995931 4696 scope.go:117] "RemoveContainer" containerID="fc5760a599a4ed277b9316717ddea596cefa13aa29e345f16dcaf3563a16f441" Nov 25 10:49:42 crc kubenswrapper[4696]: I1125 10:49:42.995503 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xcz6c" Nov 25 10:49:43 crc kubenswrapper[4696]: I1125 10:49:43.026850 4696 scope.go:117] "RemoveContainer" containerID="f66a8469d77c6d6c96cea4cee07d8676399f390e94fd032c7094560b2b524ddd" Nov 25 10:49:43 crc kubenswrapper[4696]: I1125 10:49:43.033855 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xcz6c"] Nov 25 10:49:43 crc kubenswrapper[4696]: I1125 10:49:43.037946 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xcz6c"] Nov 25 10:49:43 crc kubenswrapper[4696]: I1125 10:49:43.054313 4696 scope.go:117] "RemoveContainer" containerID="75028753c458aeb167becbd619603127cefa98daa51e42e6a36fe4d567ee7e61" Nov 25 10:49:43 crc kubenswrapper[4696]: I1125 10:49:43.391578 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b88d4b6cb-m8sz5" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.055218 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" path="/var/lib/kubelet/pods/da5dae3b-c7d7-4236-bf10-7802551338dc/volumes" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.884117 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5wkxb"] Nov 25 10:49:44 crc kubenswrapper[4696]: E1125 10:49:44.884358 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="extract-content" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.884370 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="extract-content" Nov 25 10:49:44 crc kubenswrapper[4696]: E1125 10:49:44.884387 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="extract-utilities" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.884396 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="extract-utilities" Nov 25 10:49:44 crc kubenswrapper[4696]: E1125 10:49:44.884414 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="registry-server" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.884419 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="registry-server" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.884515 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="da5dae3b-c7d7-4236-bf10-7802551338dc" containerName="registry-server" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.885256 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:44 crc kubenswrapper[4696]: I1125 10:49:44.908590 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5wkxb"] Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.031074 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-catalog-content\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.031147 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l67vg\" (UniqueName: \"kubernetes.io/projected/8555b76d-551b-43c0-93a1-aea7c941bf8a-kube-api-access-l67vg\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.031199 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-utilities\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.132940 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l67vg\" (UniqueName: \"kubernetes.io/projected/8555b76d-551b-43c0-93a1-aea7c941bf8a-kube-api-access-l67vg\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.132989 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-utilities\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.133050 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-catalog-content\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.133551 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-utilities\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.133611 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-catalog-content\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.160018 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l67vg\" (UniqueName: \"kubernetes.io/projected/8555b76d-551b-43c0-93a1-aea7c941bf8a-kube-api-access-l67vg\") pod \"certified-operators-5wkxb\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.202069 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.499814 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5wkxb"] Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.618231 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.618343 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:45 crc kubenswrapper[4696]: I1125 10:49:45.678275 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:46 crc kubenswrapper[4696]: I1125 10:49:46.018042 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wkxb" event={"ID":"8555b76d-551b-43c0-93a1-aea7c941bf8a","Type":"ContainerStarted","Data":"ff30dfa3701f2f0d624b7a30c3eaea5ceb7ea109d94a1fd27642236c01749609"} Nov 25 10:49:46 crc kubenswrapper[4696]: I1125 10:49:46.087274 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:47 crc kubenswrapper[4696]: I1125 10:49:47.027170 4696 generic.go:334] "Generic (PLEG): container finished" podID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerID="f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d" exitCode=0 Nov 25 10:49:47 crc kubenswrapper[4696]: I1125 10:49:47.027219 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wkxb" event={"ID":"8555b76d-551b-43c0-93a1-aea7c941bf8a","Type":"ContainerDied","Data":"f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d"} Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.276444 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9v4x4"] Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.487071 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l582b"] Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.488084 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.496858 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l582b"] Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.582994 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsf45\" (UniqueName: \"kubernetes.io/projected/028f9118-fa4c-47d0-8832-ffb98bbf35ee-kube-api-access-wsf45\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.583085 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-catalog-content\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.583105 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-utilities\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.684320 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsf45\" (UniqueName: \"kubernetes.io/projected/028f9118-fa4c-47d0-8832-ffb98bbf35ee-kube-api-access-wsf45\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.684362 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-catalog-content\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.684381 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-utilities\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.684825 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-catalog-content\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.684878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-utilities\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.706654 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsf45\" (UniqueName: \"kubernetes.io/projected/028f9118-fa4c-47d0-8832-ffb98bbf35ee-kube-api-access-wsf45\") pod \"redhat-operators-l582b\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:48 crc kubenswrapper[4696]: I1125 10:49:48.801531 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.038246 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9v4x4" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="registry-server" containerID="cri-o://e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6" gracePeriod=2 Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.262017 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l582b"] Nov 25 10:49:49 crc kubenswrapper[4696]: W1125 10:49:49.288269 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod028f9118_fa4c_47d0_8832_ffb98bbf35ee.slice/crio-0060a0132f790d81524b41e9c3549be824181587309bb2f3a9fa2494e8535da6 WatchSource:0}: Error finding container 0060a0132f790d81524b41e9c3549be824181587309bb2f3a9fa2494e8535da6: Status 404 returned error can't find the container with id 0060a0132f790d81524b41e9c3549be824181587309bb2f3a9fa2494e8535da6 Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.526354 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.599150 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-utilities\") pod \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.599437 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdghr\" (UniqueName: \"kubernetes.io/projected/1ea618ec-6350-4ea6-8292-d11b54f87c1c-kube-api-access-hdghr\") pod \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.599542 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-catalog-content\") pod \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\" (UID: \"1ea618ec-6350-4ea6-8292-d11b54f87c1c\") " Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.600271 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-utilities" (OuterVolumeSpecName: "utilities") pod "1ea618ec-6350-4ea6-8292-d11b54f87c1c" (UID: "1ea618ec-6350-4ea6-8292-d11b54f87c1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.605623 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea618ec-6350-4ea6-8292-d11b54f87c1c-kube-api-access-hdghr" (OuterVolumeSpecName: "kube-api-access-hdghr") pod "1ea618ec-6350-4ea6-8292-d11b54f87c1c" (UID: "1ea618ec-6350-4ea6-8292-d11b54f87c1c"). InnerVolumeSpecName "kube-api-access-hdghr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.606193 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.606232 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdghr\" (UniqueName: \"kubernetes.io/projected/1ea618ec-6350-4ea6-8292-d11b54f87c1c-kube-api-access-hdghr\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.621236 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ea618ec-6350-4ea6-8292-d11b54f87c1c" (UID: "1ea618ec-6350-4ea6-8292-d11b54f87c1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:49 crc kubenswrapper[4696]: I1125 10:49:49.712724 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea618ec-6350-4ea6-8292-d11b54f87c1c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.044485 4696 generic.go:334] "Generic (PLEG): container finished" podID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerID="4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee" exitCode=0 Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.049526 4696 generic.go:334] "Generic (PLEG): container finished" podID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerID="e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6" exitCode=0 Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.049682 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9v4x4" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.050806 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerDied","Data":"4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee"} Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.050871 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerStarted","Data":"0060a0132f790d81524b41e9c3549be824181587309bb2f3a9fa2494e8535da6"} Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.050887 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9v4x4" event={"ID":"1ea618ec-6350-4ea6-8292-d11b54f87c1c","Type":"ContainerDied","Data":"e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6"} Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.050923 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9v4x4" event={"ID":"1ea618ec-6350-4ea6-8292-d11b54f87c1c","Type":"ContainerDied","Data":"bd7c025794ddee9ed34c0c7f90d7b7ffc9d3ffe306815ad77f191a79bbd62034"} Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.050945 4696 scope.go:117] "RemoveContainer" containerID="e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.059252 4696 generic.go:334] "Generic (PLEG): container finished" podID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerID="76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395" exitCode=0 Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.059295 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wkxb" event={"ID":"8555b76d-551b-43c0-93a1-aea7c941bf8a","Type":"ContainerDied","Data":"76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395"} Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.080093 4696 scope.go:117] "RemoveContainer" containerID="e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.091824 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9v4x4"] Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.100130 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9v4x4"] Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.112410 4696 scope.go:117] "RemoveContainer" containerID="d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.124986 4696 scope.go:117] "RemoveContainer" containerID="e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6" Nov 25 10:49:50 crc kubenswrapper[4696]: E1125 10:49:50.125736 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6\": container with ID starting with e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6 not found: ID does not exist" containerID="e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.125776 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6"} err="failed to get container status \"e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6\": rpc error: code = NotFound desc = could not find container \"e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6\": container with ID starting with e29dbae721dbaa91b774ee80156595fbbd4944ed06625594746f1ae105a7cad6 not found: ID does not exist" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.125821 4696 scope.go:117] "RemoveContainer" containerID="e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61" Nov 25 10:49:50 crc kubenswrapper[4696]: E1125 10:49:50.126270 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61\": container with ID starting with e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61 not found: ID does not exist" containerID="e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.126316 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61"} err="failed to get container status \"e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61\": rpc error: code = NotFound desc = could not find container \"e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61\": container with ID starting with e10bcfe162a39259ccbfb1f88c08190ebf2a55558739d428c2080822f7b04e61 not found: ID does not exist" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.126335 4696 scope.go:117] "RemoveContainer" containerID="d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72" Nov 25 10:49:50 crc kubenswrapper[4696]: E1125 10:49:50.126642 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72\": container with ID starting with d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72 not found: ID does not exist" containerID="d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72" Nov 25 10:49:50 crc kubenswrapper[4696]: I1125 10:49:50.126743 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72"} err="failed to get container status \"d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72\": rpc error: code = NotFound desc = could not find container \"d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72\": container with ID starting with d9d624eeab5fcfb34c00a920fd9117a83079a3c8be92fdd6ebf879a6a4402e72 not found: ID does not exist" Nov 25 10:49:51 crc kubenswrapper[4696]: I1125 10:49:51.067862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wkxb" event={"ID":"8555b76d-551b-43c0-93a1-aea7c941bf8a","Type":"ContainerStarted","Data":"f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec"} Nov 25 10:49:51 crc kubenswrapper[4696]: I1125 10:49:51.069771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerStarted","Data":"6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c"} Nov 25 10:49:51 crc kubenswrapper[4696]: I1125 10:49:51.089135 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5wkxb" podStartSLOduration=3.658175337 podStartE2EDuration="7.08911974s" podCreationTimestamp="2025-11-25 10:49:44 +0000 UTC" firstStartedPulling="2025-11-25 10:49:47.029082644 +0000 UTC m=+783.842699231" lastFinishedPulling="2025-11-25 10:49:50.460027047 +0000 UTC m=+787.273643634" observedRunningTime="2025-11-25 10:49:51.088046947 +0000 UTC m=+787.901663534" watchObservedRunningTime="2025-11-25 10:49:51.08911974 +0000 UTC m=+787.902736327" Nov 25 10:49:52 crc kubenswrapper[4696]: I1125 10:49:52.050467 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" path="/var/lib/kubelet/pods/1ea618ec-6350-4ea6-8292-d11b54f87c1c/volumes" Nov 25 10:49:52 crc kubenswrapper[4696]: I1125 10:49:52.077348 4696 generic.go:334] "Generic (PLEG): container finished" podID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerID="6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c" exitCode=0 Nov 25 10:49:52 crc kubenswrapper[4696]: I1125 10:49:52.077517 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerDied","Data":"6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c"} Nov 25 10:49:53 crc kubenswrapper[4696]: I1125 10:49:53.092361 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerStarted","Data":"c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5"} Nov 25 10:49:53 crc kubenswrapper[4696]: I1125 10:49:53.119098 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l582b" podStartSLOduration=2.6803576700000002 podStartE2EDuration="5.119075016s" podCreationTimestamp="2025-11-25 10:49:48 +0000 UTC" firstStartedPulling="2025-11-25 10:49:50.046499077 +0000 UTC m=+786.860115664" lastFinishedPulling="2025-11-25 10:49:52.485216423 +0000 UTC m=+789.298833010" observedRunningTime="2025-11-25 10:49:53.117560749 +0000 UTC m=+789.931177416" watchObservedRunningTime="2025-11-25 10:49:53.119075016 +0000 UTC m=+789.932691673" Nov 25 10:49:55 crc kubenswrapper[4696]: I1125 10:49:55.203910 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:55 crc kubenswrapper[4696]: I1125 10:49:55.204049 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:55 crc kubenswrapper[4696]: I1125 10:49:55.243018 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:56 crc kubenswrapper[4696]: I1125 10:49:56.149871 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:56 crc kubenswrapper[4696]: I1125 10:49:56.876040 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5wkxb"] Nov 25 10:49:58 crc kubenswrapper[4696]: I1125 10:49:58.120303 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5wkxb" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="registry-server" containerID="cri-o://f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec" gracePeriod=2 Nov 25 10:49:58 crc kubenswrapper[4696]: I1125 10:49:58.802451 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:58 crc kubenswrapper[4696]: I1125 10:49:58.802752 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:58 crc kubenswrapper[4696]: I1125 10:49:58.847340 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.059531 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.127229 4696 generic.go:334] "Generic (PLEG): container finished" podID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerID="f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec" exitCode=0 Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.127438 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wkxb" event={"ID":"8555b76d-551b-43c0-93a1-aea7c941bf8a","Type":"ContainerDied","Data":"f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec"} Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.128359 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wkxb" event={"ID":"8555b76d-551b-43c0-93a1-aea7c941bf8a","Type":"ContainerDied","Data":"ff30dfa3701f2f0d624b7a30c3eaea5ceb7ea109d94a1fd27642236c01749609"} Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.128383 4696 scope.go:117] "RemoveContainer" containerID="f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.127524 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wkxb" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.138417 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-catalog-content\") pod \"8555b76d-551b-43c0-93a1-aea7c941bf8a\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.138657 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l67vg\" (UniqueName: \"kubernetes.io/projected/8555b76d-551b-43c0-93a1-aea7c941bf8a-kube-api-access-l67vg\") pod \"8555b76d-551b-43c0-93a1-aea7c941bf8a\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.138811 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-utilities\") pod \"8555b76d-551b-43c0-93a1-aea7c941bf8a\" (UID: \"8555b76d-551b-43c0-93a1-aea7c941bf8a\") " Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.141032 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-utilities" (OuterVolumeSpecName: "utilities") pod "8555b76d-551b-43c0-93a1-aea7c941bf8a" (UID: "8555b76d-551b-43c0-93a1-aea7c941bf8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.149152 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8555b76d-551b-43c0-93a1-aea7c941bf8a-kube-api-access-l67vg" (OuterVolumeSpecName: "kube-api-access-l67vg") pod "8555b76d-551b-43c0-93a1-aea7c941bf8a" (UID: "8555b76d-551b-43c0-93a1-aea7c941bf8a"). InnerVolumeSpecName "kube-api-access-l67vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.151912 4696 scope.go:117] "RemoveContainer" containerID="76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.178330 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.181493 4696 scope.go:117] "RemoveContainer" containerID="f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.198714 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8555b76d-551b-43c0-93a1-aea7c941bf8a" (UID: "8555b76d-551b-43c0-93a1-aea7c941bf8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.199894 4696 scope.go:117] "RemoveContainer" containerID="f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec" Nov 25 10:49:59 crc kubenswrapper[4696]: E1125 10:49:59.200812 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec\": container with ID starting with f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec not found: ID does not exist" containerID="f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.200918 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec"} err="failed to get container status \"f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec\": rpc error: code = NotFound desc = could not find container \"f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec\": container with ID starting with f8dafeaa9733e0e9a1c21b808dfeceaaadf97bdf91e7aae27428be8637647dec not found: ID does not exist" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.201004 4696 scope.go:117] "RemoveContainer" containerID="76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395" Nov 25 10:49:59 crc kubenswrapper[4696]: E1125 10:49:59.201997 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395\": container with ID starting with 76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395 not found: ID does not exist" containerID="76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.202019 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395"} err="failed to get container status \"76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395\": rpc error: code = NotFound desc = could not find container \"76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395\": container with ID starting with 76fd236d14b1755afdeb2e63181848ab854f801536cf72efd47f9c83a0456395 not found: ID does not exist" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.202034 4696 scope.go:117] "RemoveContainer" containerID="f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d" Nov 25 10:49:59 crc kubenswrapper[4696]: E1125 10:49:59.202225 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d\": container with ID starting with f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d not found: ID does not exist" containerID="f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.202308 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d"} err="failed to get container status \"f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d\": rpc error: code = NotFound desc = could not find container \"f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d\": container with ID starting with f7a5a80cd2c4f0ed36e678ed2cf5c12ea86ae33636b91faa6d653d3cb49c567d not found: ID does not exist" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.240903 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.240935 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l67vg\" (UniqueName: \"kubernetes.io/projected/8555b76d-551b-43c0-93a1-aea7c941bf8a-kube-api-access-l67vg\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.240950 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8555b76d-551b-43c0-93a1-aea7c941bf8a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.455542 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5wkxb"] Nov 25 10:49:59 crc kubenswrapper[4696]: I1125 10:49:59.460721 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5wkxb"] Nov 25 10:50:00 crc kubenswrapper[4696]: I1125 10:50:00.053607 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" path="/var/lib/kubelet/pods/8555b76d-551b-43c0-93a1-aea7c941bf8a/volumes" Nov 25 10:50:00 crc kubenswrapper[4696]: I1125 10:50:00.801587 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:50:00 crc kubenswrapper[4696]: I1125 10:50:00.801649 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:50:01 crc kubenswrapper[4696]: I1125 10:50:01.677212 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l582b"] Nov 25 10:50:01 crc kubenswrapper[4696]: I1125 10:50:01.677403 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l582b" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="registry-server" containerID="cri-o://c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5" gracePeriod=2 Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.065214 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.148603 4696 generic.go:334] "Generic (PLEG): container finished" podID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerID="c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5" exitCode=0 Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.148648 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerDied","Data":"c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5"} Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.148702 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l582b" event={"ID":"028f9118-fa4c-47d0-8832-ffb98bbf35ee","Type":"ContainerDied","Data":"0060a0132f790d81524b41e9c3549be824181587309bb2f3a9fa2494e8535da6"} Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.148720 4696 scope.go:117] "RemoveContainer" containerID="c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.149110 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l582b" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.164755 4696 scope.go:117] "RemoveContainer" containerID="6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.177726 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-utilities\") pod \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.178382 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-catalog-content\") pod \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.178552 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsf45\" (UniqueName: \"kubernetes.io/projected/028f9118-fa4c-47d0-8832-ffb98bbf35ee-kube-api-access-wsf45\") pod \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\" (UID: \"028f9118-fa4c-47d0-8832-ffb98bbf35ee\") " Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.183770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-utilities" (OuterVolumeSpecName: "utilities") pod "028f9118-fa4c-47d0-8832-ffb98bbf35ee" (UID: "028f9118-fa4c-47d0-8832-ffb98bbf35ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.190046 4696 scope.go:117] "RemoveContainer" containerID="4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.190072 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/028f9118-fa4c-47d0-8832-ffb98bbf35ee-kube-api-access-wsf45" (OuterVolumeSpecName: "kube-api-access-wsf45") pod "028f9118-fa4c-47d0-8832-ffb98bbf35ee" (UID: "028f9118-fa4c-47d0-8832-ffb98bbf35ee"). InnerVolumeSpecName "kube-api-access-wsf45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.214759 4696 scope.go:117] "RemoveContainer" containerID="c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5" Nov 25 10:50:02 crc kubenswrapper[4696]: E1125 10:50:02.215149 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5\": container with ID starting with c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5 not found: ID does not exist" containerID="c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.215179 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5"} err="failed to get container status \"c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5\": rpc error: code = NotFound desc = could not find container \"c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5\": container with ID starting with c375467858c6855eeed0fa4a97ff7b0d89955bd3bd8d2914c6901ce20ade62f5 not found: ID does not exist" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.215200 4696 scope.go:117] "RemoveContainer" containerID="6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c" Nov 25 10:50:02 crc kubenswrapper[4696]: E1125 10:50:02.215539 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c\": container with ID starting with 6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c not found: ID does not exist" containerID="6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.215560 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c"} err="failed to get container status \"6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c\": rpc error: code = NotFound desc = could not find container \"6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c\": container with ID starting with 6df867780f463813dd0ba13537e87d77150d7940ddcf8ac9f38ebac6965b602c not found: ID does not exist" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.215572 4696 scope.go:117] "RemoveContainer" containerID="4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee" Nov 25 10:50:02 crc kubenswrapper[4696]: E1125 10:50:02.215910 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee\": container with ID starting with 4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee not found: ID does not exist" containerID="4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.215931 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee"} err="failed to get container status \"4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee\": rpc error: code = NotFound desc = could not find container \"4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee\": container with ID starting with 4b847af842584e92374789f54b45fb3fd23b3df783e26556145c90f448801eee not found: ID does not exist" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.280736 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "028f9118-fa4c-47d0-8832-ffb98bbf35ee" (UID: "028f9118-fa4c-47d0-8832-ffb98bbf35ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.281520 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsf45\" (UniqueName: \"kubernetes.io/projected/028f9118-fa4c-47d0-8832-ffb98bbf35ee-kube-api-access-wsf45\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.281560 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.281573 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028f9118-fa4c-47d0-8832-ffb98bbf35ee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.477316 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l582b"] Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.480775 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l582b"] Nov 25 10:50:02 crc kubenswrapper[4696]: I1125 10:50:02.951361 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-d5845fdbd-84v4v" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710512 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-nhrzx"] Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710757 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="extract-utilities" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710769 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="extract-utilities" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710779 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710785 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710794 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="extract-content" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710800 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="extract-content" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710809 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710815 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710824 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="extract-content" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710830 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="extract-content" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710839 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="extract-utilities" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710844 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="extract-utilities" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710853 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710859 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710867 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="extract-utilities" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710872 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="extract-utilities" Nov 25 10:50:03 crc kubenswrapper[4696]: E1125 10:50:03.710882 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="extract-content" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710889 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="extract-content" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.710997 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea618ec-6350-4ea6-8292-d11b54f87c1c" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.711005 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8555b76d-551b-43c0-93a1-aea7c941bf8a" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.711018 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" containerName="registry-server" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.712657 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.716358 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.716701 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.716817 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4n87h" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.725686 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6"] Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.729042 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.730929 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.739300 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6"] Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.804457 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-metrics\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.804967 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1462b48a-8222-418b-bcb1-06b9d3a2f99b-metrics-certs\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.805228 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-startup\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.805278 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-conf\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.805309 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-reloader\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.805394 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtq8v\" (UniqueName: \"kubernetes.io/projected/1462b48a-8222-418b-bcb1-06b9d3a2f99b-kube-api-access-gtq8v\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.805452 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-sockets\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.837096 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-42dc9"] Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.838099 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-42dc9" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.839818 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2qwbw" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.839953 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.840252 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.840253 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.873632 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-jgmgn"] Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.874722 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.877189 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.881459 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-jgmgn"] Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906406 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-startup\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-conf\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906462 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-reloader\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906486 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5lwr\" (UniqueName: \"kubernetes.io/projected/c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f-kube-api-access-k5lwr\") pod \"frr-k8s-webhook-server-6998585d5-nd8g6\" (UID: \"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906511 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtq8v\" (UniqueName: \"kubernetes.io/projected/1462b48a-8222-418b-bcb1-06b9d3a2f99b-kube-api-access-gtq8v\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906526 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f-cert\") pod \"frr-k8s-webhook-server-6998585d5-nd8g6\" (UID: \"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906551 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-sockets\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906578 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-metrics\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.906599 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1462b48a-8222-418b-bcb1-06b9d3a2f99b-metrics-certs\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.907508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-metrics\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.907508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-conf\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.907590 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-reloader\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.907829 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-sockets\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.908099 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1462b48a-8222-418b-bcb1-06b9d3a2f99b-frr-startup\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.911050 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1462b48a-8222-418b-bcb1-06b9d3a2f99b-metrics-certs\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:03 crc kubenswrapper[4696]: I1125 10:50:03.953000 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtq8v\" (UniqueName: \"kubernetes.io/projected/1462b48a-8222-418b-bcb1-06b9d3a2f99b-kube-api-access-gtq8v\") pod \"frr-k8s-nhrzx\" (UID: \"1462b48a-8222-418b-bcb1-06b9d3a2f99b\") " pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007523 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-metrics-certs\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007566 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007588 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e4999207-9162-4635-b7a8-2bd0d5311019-metallb-excludel2\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007612 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8twgc\" (UniqueName: \"kubernetes.io/projected/e4999207-9162-4635-b7a8-2bd0d5311019-kube-api-access-8twgc\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007635 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckk64\" (UniqueName: \"kubernetes.io/projected/b24625b7-df3d-4f50-ba97-5eac956b1662-kube-api-access-ckk64\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007658 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-metrics-certs\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007698 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5lwr\" (UniqueName: \"kubernetes.io/projected/c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f-kube-api-access-k5lwr\") pod \"frr-k8s-webhook-server-6998585d5-nd8g6\" (UID: \"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007720 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f-cert\") pod \"frr-k8s-webhook-server-6998585d5-nd8g6\" (UID: \"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.007745 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-cert\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.012261 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f-cert\") pod \"frr-k8s-webhook-server-6998585d5-nd8g6\" (UID: \"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.036333 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5lwr\" (UniqueName: \"kubernetes.io/projected/c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f-kube-api-access-k5lwr\") pod \"frr-k8s-webhook-server-6998585d5-nd8g6\" (UID: \"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.039876 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.063000 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="028f9118-fa4c-47d0-8832-ffb98bbf35ee" path="/var/lib/kubelet/pods/028f9118-fa4c-47d0-8832-ffb98bbf35ee/volumes" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.075294 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110210 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckk64\" (UniqueName: \"kubernetes.io/projected/b24625b7-df3d-4f50-ba97-5eac956b1662-kube-api-access-ckk64\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110263 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-metrics-certs\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110306 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-cert\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110337 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-metrics-certs\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110353 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110371 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e4999207-9162-4635-b7a8-2bd0d5311019-metallb-excludel2\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.110390 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8twgc\" (UniqueName: \"kubernetes.io/projected/e4999207-9162-4635-b7a8-2bd0d5311019-kube-api-access-8twgc\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: E1125 10:50:04.110720 4696 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 25 10:50:04 crc kubenswrapper[4696]: E1125 10:50:04.110764 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-metrics-certs podName:b24625b7-df3d-4f50-ba97-5eac956b1662 nodeName:}" failed. No retries permitted until 2025-11-25 10:50:04.610749936 +0000 UTC m=+801.424366523 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-metrics-certs") pod "controller-6c7b4b5f48-jgmgn" (UID: "b24625b7-df3d-4f50-ba97-5eac956b1662") : secret "controller-certs-secret" not found Nov 25 10:50:04 crc kubenswrapper[4696]: E1125 10:50:04.110908 4696 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 10:50:04 crc kubenswrapper[4696]: E1125 10:50:04.110935 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist podName:e4999207-9162-4635-b7a8-2bd0d5311019 nodeName:}" failed. No retries permitted until 2025-11-25 10:50:04.610928372 +0000 UTC m=+801.424544959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist") pod "speaker-42dc9" (UID: "e4999207-9162-4635-b7a8-2bd0d5311019") : secret "metallb-memberlist" not found Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.113498 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e4999207-9162-4635-b7a8-2bd0d5311019-metallb-excludel2\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.113848 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.117335 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-metrics-certs\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.130110 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-cert\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.130459 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8twgc\" (UniqueName: \"kubernetes.io/projected/e4999207-9162-4635-b7a8-2bd0d5311019-kube-api-access-8twgc\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.136361 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckk64\" (UniqueName: \"kubernetes.io/projected/b24625b7-df3d-4f50-ba97-5eac956b1662-kube-api-access-ckk64\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.520190 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6"] Nov 25 10:50:04 crc kubenswrapper[4696]: W1125 10:50:04.522734 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2b796fe_37ae_4c76_b1a0_fe8e916f8c6f.slice/crio-58e531a032bd9a18024e75cb60c1b3ac95be27db3b7a57f40c6105c87a395683 WatchSource:0}: Error finding container 58e531a032bd9a18024e75cb60c1b3ac95be27db3b7a57f40c6105c87a395683: Status 404 returned error can't find the container with id 58e531a032bd9a18024e75cb60c1b3ac95be27db3b7a57f40c6105c87a395683 Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.619051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.619214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-metrics-certs\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: E1125 10:50:04.619246 4696 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 10:50:04 crc kubenswrapper[4696]: E1125 10:50:04.619324 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist podName:e4999207-9162-4635-b7a8-2bd0d5311019 nodeName:}" failed. No retries permitted until 2025-11-25 10:50:05.619307801 +0000 UTC m=+802.432924388 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist") pod "speaker-42dc9" (UID: "e4999207-9162-4635-b7a8-2bd0d5311019") : secret "metallb-memberlist" not found Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.624696 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b24625b7-df3d-4f50-ba97-5eac956b1662-metrics-certs\") pod \"controller-6c7b4b5f48-jgmgn\" (UID: \"b24625b7-df3d-4f50-ba97-5eac956b1662\") " pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.790933 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:04 crc kubenswrapper[4696]: I1125 10:50:04.987547 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-jgmgn"] Nov 25 10:50:04 crc kubenswrapper[4696]: W1125 10:50:04.996606 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb24625b7_df3d_4f50_ba97_5eac956b1662.slice/crio-7afc4f5461b286eea2ba48e10fe0ccb99251fa033882ce861e9c1a79482219aa WatchSource:0}: Error finding container 7afc4f5461b286eea2ba48e10fe0ccb99251fa033882ce861e9c1a79482219aa: Status 404 returned error can't find the container with id 7afc4f5461b286eea2ba48e10fe0ccb99251fa033882ce861e9c1a79482219aa Nov 25 10:50:05 crc kubenswrapper[4696]: I1125 10:50:05.166256 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"e4219984867863a211d0477f305b20a4d16aed3f4b30da30b613a687d61ce1a9"} Nov 25 10:50:05 crc kubenswrapper[4696]: I1125 10:50:05.167289 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-jgmgn" event={"ID":"b24625b7-df3d-4f50-ba97-5eac956b1662","Type":"ContainerStarted","Data":"7afc4f5461b286eea2ba48e10fe0ccb99251fa033882ce861e9c1a79482219aa"} Nov 25 10:50:05 crc kubenswrapper[4696]: I1125 10:50:05.168260 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" event={"ID":"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f","Type":"ContainerStarted","Data":"58e531a032bd9a18024e75cb60c1b3ac95be27db3b7a57f40c6105c87a395683"} Nov 25 10:50:05 crc kubenswrapper[4696]: I1125 10:50:05.634715 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:05 crc kubenswrapper[4696]: I1125 10:50:05.642326 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e4999207-9162-4635-b7a8-2bd0d5311019-memberlist\") pod \"speaker-42dc9\" (UID: \"e4999207-9162-4635-b7a8-2bd0d5311019\") " pod="metallb-system/speaker-42dc9" Nov 25 10:50:05 crc kubenswrapper[4696]: I1125 10:50:05.659363 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-42dc9" Nov 25 10:50:06 crc kubenswrapper[4696]: I1125 10:50:06.176357 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-42dc9" event={"ID":"e4999207-9162-4635-b7a8-2bd0d5311019","Type":"ContainerStarted","Data":"5da6c8968969ce0c89290eecc7731be9c21afea7f71f9efd451de0a282a584e8"} Nov 25 10:50:06 crc kubenswrapper[4696]: I1125 10:50:06.176587 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-42dc9" event={"ID":"e4999207-9162-4635-b7a8-2bd0d5311019","Type":"ContainerStarted","Data":"e6e7dea85bd172767b6fc1f1f3b7e48b438c312824028011a81f98fa7442af9b"} Nov 25 10:50:06 crc kubenswrapper[4696]: I1125 10:50:06.179504 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-jgmgn" event={"ID":"b24625b7-df3d-4f50-ba97-5eac956b1662","Type":"ContainerStarted","Data":"fd0c2cc303b02ced8f90465722885375db08a2154cdbe988b026c151a4ba7158"} Nov 25 10:50:06 crc kubenswrapper[4696]: I1125 10:50:06.179539 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-jgmgn" event={"ID":"b24625b7-df3d-4f50-ba97-5eac956b1662","Type":"ContainerStarted","Data":"50723408ac3df65a3da5f9e4d0d5a2c6d9696e9dbb851e9509e6083adc961a2c"} Nov 25 10:50:06 crc kubenswrapper[4696]: I1125 10:50:06.179775 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:06 crc kubenswrapper[4696]: I1125 10:50:06.200845 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-jgmgn" podStartSLOduration=3.200828818 podStartE2EDuration="3.200828818s" podCreationTimestamp="2025-11-25 10:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:06.1987006 +0000 UTC m=+803.012317177" watchObservedRunningTime="2025-11-25 10:50:06.200828818 +0000 UTC m=+803.014445405" Nov 25 10:50:07 crc kubenswrapper[4696]: I1125 10:50:07.191253 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-42dc9" event={"ID":"e4999207-9162-4635-b7a8-2bd0d5311019","Type":"ContainerStarted","Data":"da07504d96af5eae18abb2b34dc1dfd231f81c56cd1919cb58d053a768757b99"} Nov 25 10:50:08 crc kubenswrapper[4696]: I1125 10:50:08.212518 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-42dc9" Nov 25 10:50:12 crc kubenswrapper[4696]: I1125 10:50:12.234215 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" event={"ID":"c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f","Type":"ContainerStarted","Data":"b7536c72a4afb287421c1d38a77e4246a40c90a7a1b3675b908aa6a99d481c0b"} Nov 25 10:50:12 crc kubenswrapper[4696]: I1125 10:50:12.234617 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:12 crc kubenswrapper[4696]: I1125 10:50:12.236759 4696 generic.go:334] "Generic (PLEG): container finished" podID="1462b48a-8222-418b-bcb1-06b9d3a2f99b" containerID="9b5f5fa1f74bf0376f60e59912d561bcfa61578c9cb5bf19075722dfbaf2b782" exitCode=0 Nov 25 10:50:12 crc kubenswrapper[4696]: I1125 10:50:12.236807 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerDied","Data":"9b5f5fa1f74bf0376f60e59912d561bcfa61578c9cb5bf19075722dfbaf2b782"} Nov 25 10:50:12 crc kubenswrapper[4696]: I1125 10:50:12.252656 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-42dc9" podStartSLOduration=9.252639537 podStartE2EDuration="9.252639537s" podCreationTimestamp="2025-11-25 10:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:07.250111742 +0000 UTC m=+804.063728349" watchObservedRunningTime="2025-11-25 10:50:12.252639537 +0000 UTC m=+809.066256124" Nov 25 10:50:12 crc kubenswrapper[4696]: I1125 10:50:12.254037 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" podStartSLOduration=1.891138197 podStartE2EDuration="9.25403279s" podCreationTimestamp="2025-11-25 10:50:03 +0000 UTC" firstStartedPulling="2025-11-25 10:50:04.524823804 +0000 UTC m=+801.338440391" lastFinishedPulling="2025-11-25 10:50:11.887718397 +0000 UTC m=+808.701334984" observedRunningTime="2025-11-25 10:50:12.248086163 +0000 UTC m=+809.061702750" watchObservedRunningTime="2025-11-25 10:50:12.25403279 +0000 UTC m=+809.067649377" Nov 25 10:50:13 crc kubenswrapper[4696]: I1125 10:50:13.244104 4696 generic.go:334] "Generic (PLEG): container finished" podID="1462b48a-8222-418b-bcb1-06b9d3a2f99b" containerID="7ef8dbe492a768117acda914bc12a5a20ec8afc99e1c42fcb71eaf1ceb6a6844" exitCode=0 Nov 25 10:50:13 crc kubenswrapper[4696]: I1125 10:50:13.244218 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerDied","Data":"7ef8dbe492a768117acda914bc12a5a20ec8afc99e1c42fcb71eaf1ceb6a6844"} Nov 25 10:50:14 crc kubenswrapper[4696]: I1125 10:50:14.251488 4696 generic.go:334] "Generic (PLEG): container finished" podID="1462b48a-8222-418b-bcb1-06b9d3a2f99b" containerID="21acab6b2b735238a8c9e8655dac490383bcf142778c3dd9d453b8449c16db5c" exitCode=0 Nov 25 10:50:14 crc kubenswrapper[4696]: I1125 10:50:14.251812 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerDied","Data":"21acab6b2b735238a8c9e8655dac490383bcf142778c3dd9d453b8449c16db5c"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.261525 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"08759b3df4dcea088a91bd7889dca0a2dfc23d1b22bf56c6325d277a04a5d6f5"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.262789 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.262808 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"952a9dc2b516b109425696e4140111f053d1450a394c3817093e3a1e130b3013"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.262820 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"8d32722d63b1150bd35fc46c069db70aff9ac036a9344f4b55f10c5cf34afb3f"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.262828 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"1bd99dcec58f47bfeaf876b5d67daaac14f6ce20ea376b50296656fbc1bdfa57"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.262837 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"26cc8b9e2a8a391a1b8c866171434f489343f21cb323e5695785b2656ac78127"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.262844 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhrzx" event={"ID":"1462b48a-8222-418b-bcb1-06b9d3a2f99b","Type":"ContainerStarted","Data":"95bc71bc534db8e608744598925960b60ed433312eaeb957efc6f9b364e627cd"} Nov 25 10:50:15 crc kubenswrapper[4696]: I1125 10:50:15.286988 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-nhrzx" podStartSLOduration=4.620534974 podStartE2EDuration="12.286968012s" podCreationTimestamp="2025-11-25 10:50:03 +0000 UTC" firstStartedPulling="2025-11-25 10:50:04.19751236 +0000 UTC m=+801.011128947" lastFinishedPulling="2025-11-25 10:50:11.863945398 +0000 UTC m=+808.677561985" observedRunningTime="2025-11-25 10:50:15.283247164 +0000 UTC m=+812.096863771" watchObservedRunningTime="2025-11-25 10:50:15.286968012 +0000 UTC m=+812.100584589" Nov 25 10:50:19 crc kubenswrapper[4696]: I1125 10:50:19.040942 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:19 crc kubenswrapper[4696]: I1125 10:50:19.092519 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:24 crc kubenswrapper[4696]: I1125 10:50:24.051387 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-nhrzx" Nov 25 10:50:24 crc kubenswrapper[4696]: I1125 10:50:24.082297 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-nd8g6" Nov 25 10:50:24 crc kubenswrapper[4696]: I1125 10:50:24.795537 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-jgmgn" Nov 25 10:50:25 crc kubenswrapper[4696]: I1125 10:50:25.664032 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-42dc9" Nov 25 10:50:28 crc kubenswrapper[4696]: I1125 10:50:28.906272 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7p4w7"] Nov 25 10:50:28 crc kubenswrapper[4696]: I1125 10:50:28.907186 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:28 crc kubenswrapper[4696]: I1125 10:50:28.908985 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-m844t" Nov 25 10:50:28 crc kubenswrapper[4696]: I1125 10:50:28.915451 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 10:50:28 crc kubenswrapper[4696]: I1125 10:50:28.916024 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 10:50:28 crc kubenswrapper[4696]: I1125 10:50:28.955223 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7p4w7"] Nov 25 10:50:29 crc kubenswrapper[4696]: I1125 10:50:29.057496 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz66l\" (UniqueName: \"kubernetes.io/projected/8462d2f1-3a40-44fa-83c2-e71ec0c52765-kube-api-access-fz66l\") pod \"openstack-operator-index-7p4w7\" (UID: \"8462d2f1-3a40-44fa-83c2-e71ec0c52765\") " pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:29 crc kubenswrapper[4696]: I1125 10:50:29.158795 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz66l\" (UniqueName: \"kubernetes.io/projected/8462d2f1-3a40-44fa-83c2-e71ec0c52765-kube-api-access-fz66l\") pod \"openstack-operator-index-7p4w7\" (UID: \"8462d2f1-3a40-44fa-83c2-e71ec0c52765\") " pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:29 crc kubenswrapper[4696]: I1125 10:50:29.178882 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz66l\" (UniqueName: \"kubernetes.io/projected/8462d2f1-3a40-44fa-83c2-e71ec0c52765-kube-api-access-fz66l\") pod \"openstack-operator-index-7p4w7\" (UID: \"8462d2f1-3a40-44fa-83c2-e71ec0c52765\") " pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:29 crc kubenswrapper[4696]: I1125 10:50:29.226870 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:29 crc kubenswrapper[4696]: I1125 10:50:29.623110 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7p4w7"] Nov 25 10:50:29 crc kubenswrapper[4696]: W1125 10:50:29.627843 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8462d2f1_3a40_44fa_83c2_e71ec0c52765.slice/crio-fb47349fd94561b612a1c456a0908a4ef7aa98801495da7a51557e8c7c926ec3 WatchSource:0}: Error finding container fb47349fd94561b612a1c456a0908a4ef7aa98801495da7a51557e8c7c926ec3: Status 404 returned error can't find the container with id fb47349fd94561b612a1c456a0908a4ef7aa98801495da7a51557e8c7c926ec3 Nov 25 10:50:30 crc kubenswrapper[4696]: I1125 10:50:30.357719 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p4w7" event={"ID":"8462d2f1-3a40-44fa-83c2-e71ec0c52765","Type":"ContainerStarted","Data":"fb47349fd94561b612a1c456a0908a4ef7aa98801495da7a51557e8c7c926ec3"} Nov 25 10:50:30 crc kubenswrapper[4696]: I1125 10:50:30.802084 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:50:30 crc kubenswrapper[4696]: I1125 10:50:30.802436 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:50:31 crc kubenswrapper[4696]: I1125 10:50:31.877711 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7p4w7"] Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.488908 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-55w9h"] Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.490372 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.494451 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-55w9h"] Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.606696 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbgt8\" (UniqueName: \"kubernetes.io/projected/2d7b9c54-45bc-4bc8-9513-a37f199b022f-kube-api-access-lbgt8\") pod \"openstack-operator-index-55w9h\" (UID: \"2d7b9c54-45bc-4bc8-9513-a37f199b022f\") " pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.708643 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbgt8\" (UniqueName: \"kubernetes.io/projected/2d7b9c54-45bc-4bc8-9513-a37f199b022f-kube-api-access-lbgt8\") pod \"openstack-operator-index-55w9h\" (UID: \"2d7b9c54-45bc-4bc8-9513-a37f199b022f\") " pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.733781 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbgt8\" (UniqueName: \"kubernetes.io/projected/2d7b9c54-45bc-4bc8-9513-a37f199b022f-kube-api-access-lbgt8\") pod \"openstack-operator-index-55w9h\" (UID: \"2d7b9c54-45bc-4bc8-9513-a37f199b022f\") " pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:32 crc kubenswrapper[4696]: I1125 10:50:32.816499 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.254219 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-55w9h"] Nov 25 10:50:33 crc kubenswrapper[4696]: W1125 10:50:33.259565 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d7b9c54_45bc_4bc8_9513_a37f199b022f.slice/crio-60d6830ababfcdb135f40d6e3b06499b798ebe350a5add48040b392d6bdca4d3 WatchSource:0}: Error finding container 60d6830ababfcdb135f40d6e3b06499b798ebe350a5add48040b392d6bdca4d3: Status 404 returned error can't find the container with id 60d6830ababfcdb135f40d6e3b06499b798ebe350a5add48040b392d6bdca4d3 Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.376029 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p4w7" event={"ID":"8462d2f1-3a40-44fa-83c2-e71ec0c52765","Type":"ContainerStarted","Data":"18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107"} Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.376132 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7p4w7" podUID="8462d2f1-3a40-44fa-83c2-e71ec0c52765" containerName="registry-server" containerID="cri-o://18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107" gracePeriod=2 Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.380392 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-55w9h" event={"ID":"2d7b9c54-45bc-4bc8-9513-a37f199b022f","Type":"ContainerStarted","Data":"60d6830ababfcdb135f40d6e3b06499b798ebe350a5add48040b392d6bdca4d3"} Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.396081 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7p4w7" podStartSLOduration=2.166953438 podStartE2EDuration="5.39606662s" podCreationTimestamp="2025-11-25 10:50:28 +0000 UTC" firstStartedPulling="2025-11-25 10:50:29.629723709 +0000 UTC m=+826.443340286" lastFinishedPulling="2025-11-25 10:50:32.858836891 +0000 UTC m=+829.672453468" observedRunningTime="2025-11-25 10:50:33.395843174 +0000 UTC m=+830.209459761" watchObservedRunningTime="2025-11-25 10:50:33.39606662 +0000 UTC m=+830.209683207" Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.688430 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.824851 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz66l\" (UniqueName: \"kubernetes.io/projected/8462d2f1-3a40-44fa-83c2-e71ec0c52765-kube-api-access-fz66l\") pod \"8462d2f1-3a40-44fa-83c2-e71ec0c52765\" (UID: \"8462d2f1-3a40-44fa-83c2-e71ec0c52765\") " Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.831575 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8462d2f1-3a40-44fa-83c2-e71ec0c52765-kube-api-access-fz66l" (OuterVolumeSpecName: "kube-api-access-fz66l") pod "8462d2f1-3a40-44fa-83c2-e71ec0c52765" (UID: "8462d2f1-3a40-44fa-83c2-e71ec0c52765"). InnerVolumeSpecName "kube-api-access-fz66l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:33 crc kubenswrapper[4696]: I1125 10:50:33.926794 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz66l\" (UniqueName: \"kubernetes.io/projected/8462d2f1-3a40-44fa-83c2-e71ec0c52765-kube-api-access-fz66l\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:34 crc kubenswrapper[4696]: E1125 10:50:34.159368 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8462d2f1_3a40_44fa_83c2_e71ec0c52765.slice\": RecentStats: unable to find data in memory cache]" Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.387536 4696 generic.go:334] "Generic (PLEG): container finished" podID="8462d2f1-3a40-44fa-83c2-e71ec0c52765" containerID="18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107" exitCode=0 Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.387585 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7p4w7" Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.388383 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p4w7" event={"ID":"8462d2f1-3a40-44fa-83c2-e71ec0c52765","Type":"ContainerDied","Data":"18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107"} Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.388573 4696 scope.go:117] "RemoveContainer" containerID="18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107" Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.388686 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7p4w7" event={"ID":"8462d2f1-3a40-44fa-83c2-e71ec0c52765","Type":"ContainerDied","Data":"fb47349fd94561b612a1c456a0908a4ef7aa98801495da7a51557e8c7c926ec3"} Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.390325 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-55w9h" event={"ID":"2d7b9c54-45bc-4bc8-9513-a37f199b022f","Type":"ContainerStarted","Data":"d22e2ec9dfe84dc52bf2b1ce388d2cf0e433c33d15c648b3584d23b891be04bc"} Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.405296 4696 scope.go:117] "RemoveContainer" containerID="18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107" Nov 25 10:50:34 crc kubenswrapper[4696]: E1125 10:50:34.406215 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107\": container with ID starting with 18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107 not found: ID does not exist" containerID="18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107" Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.406329 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107"} err="failed to get container status \"18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107\": rpc error: code = NotFound desc = could not find container \"18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107\": container with ID starting with 18d9df446c1c4ff7aa53ced6adba8af4576d77c9f3d3e2db3f5b93cc64e72107 not found: ID does not exist" Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.409956 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-55w9h" podStartSLOduration=2.346693756 podStartE2EDuration="2.409934658s" podCreationTimestamp="2025-11-25 10:50:32 +0000 UTC" firstStartedPulling="2025-11-25 10:50:33.263110531 +0000 UTC m=+830.076727118" lastFinishedPulling="2025-11-25 10:50:33.326351433 +0000 UTC m=+830.139968020" observedRunningTime="2025-11-25 10:50:34.408650478 +0000 UTC m=+831.222267065" watchObservedRunningTime="2025-11-25 10:50:34.409934658 +0000 UTC m=+831.223551245" Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.425861 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7p4w7"] Nov 25 10:50:34 crc kubenswrapper[4696]: I1125 10:50:34.430410 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7p4w7"] Nov 25 10:50:36 crc kubenswrapper[4696]: I1125 10:50:36.050633 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8462d2f1-3a40-44fa-83c2-e71ec0c52765" path="/var/lib/kubelet/pods/8462d2f1-3a40-44fa-83c2-e71ec0c52765/volumes" Nov 25 10:50:42 crc kubenswrapper[4696]: I1125 10:50:42.817625 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:42 crc kubenswrapper[4696]: I1125 10:50:42.818205 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:42 crc kubenswrapper[4696]: I1125 10:50:42.846543 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:43 crc kubenswrapper[4696]: I1125 10:50:43.473620 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-55w9h" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.133012 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5"] Nov 25 10:50:45 crc kubenswrapper[4696]: E1125 10:50:45.133558 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8462d2f1-3a40-44fa-83c2-e71ec0c52765" containerName="registry-server" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.133572 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8462d2f1-3a40-44fa-83c2-e71ec0c52765" containerName="registry-server" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.133758 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8462d2f1-3a40-44fa-83c2-e71ec0c52765" containerName="registry-server" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.134953 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.139002 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-rbrt6" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.142866 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5"] Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.283312 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-util\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.283376 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj4vz\" (UniqueName: \"kubernetes.io/projected/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-kube-api-access-pj4vz\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.283486 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-bundle\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.384940 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-bundle\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.385038 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-util\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.385079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj4vz\" (UniqueName: \"kubernetes.io/projected/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-kube-api-access-pj4vz\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.385591 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-bundle\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.385624 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-util\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.404037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj4vz\" (UniqueName: \"kubernetes.io/projected/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-kube-api-access-pj4vz\") pod \"3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.451176 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:45 crc kubenswrapper[4696]: I1125 10:50:45.877922 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5"] Nov 25 10:50:46 crc kubenswrapper[4696]: I1125 10:50:46.472893 4696 generic.go:334] "Generic (PLEG): container finished" podID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerID="df8d0ba978f7e1a7c3cd6e89865214afecc144872d071a9a6f660cb75ac9042d" exitCode=0 Nov 25 10:50:46 crc kubenswrapper[4696]: I1125 10:50:46.473160 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" event={"ID":"e0a75471-aabe-4787-a3ec-fedcf4cecbf1","Type":"ContainerDied","Data":"df8d0ba978f7e1a7c3cd6e89865214afecc144872d071a9a6f660cb75ac9042d"} Nov 25 10:50:46 crc kubenswrapper[4696]: I1125 10:50:46.473209 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" event={"ID":"e0a75471-aabe-4787-a3ec-fedcf4cecbf1","Type":"ContainerStarted","Data":"09085930e97cfdc810700157d1aa67ff74d75433952844b07eb29a277c768792"} Nov 25 10:50:49 crc kubenswrapper[4696]: I1125 10:50:49.491215 4696 generic.go:334] "Generic (PLEG): container finished" podID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerID="d272acd7d91125582ede60de551aeacccf0f65d6c007c9d7cb9c44abbf4824bf" exitCode=0 Nov 25 10:50:49 crc kubenswrapper[4696]: I1125 10:50:49.491287 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" event={"ID":"e0a75471-aabe-4787-a3ec-fedcf4cecbf1","Type":"ContainerDied","Data":"d272acd7d91125582ede60de551aeacccf0f65d6c007c9d7cb9c44abbf4824bf"} Nov 25 10:50:50 crc kubenswrapper[4696]: I1125 10:50:50.500025 4696 generic.go:334] "Generic (PLEG): container finished" podID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerID="98e0200c95f491d4499769bf9a1d496270da196f1b5ffa20c99973931aea4f0f" exitCode=0 Nov 25 10:50:50 crc kubenswrapper[4696]: I1125 10:50:50.500324 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" event={"ID":"e0a75471-aabe-4787-a3ec-fedcf4cecbf1","Type":"ContainerDied","Data":"98e0200c95f491d4499769bf9a1d496270da196f1b5ffa20c99973931aea4f0f"} Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.754098 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.868497 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-bundle\") pod \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.868593 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-util\") pod \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.868699 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj4vz\" (UniqueName: \"kubernetes.io/projected/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-kube-api-access-pj4vz\") pod \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\" (UID: \"e0a75471-aabe-4787-a3ec-fedcf4cecbf1\") " Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.869497 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-bundle" (OuterVolumeSpecName: "bundle") pod "e0a75471-aabe-4787-a3ec-fedcf4cecbf1" (UID: "e0a75471-aabe-4787-a3ec-fedcf4cecbf1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.874910 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-kube-api-access-pj4vz" (OuterVolumeSpecName: "kube-api-access-pj4vz") pod "e0a75471-aabe-4787-a3ec-fedcf4cecbf1" (UID: "e0a75471-aabe-4787-a3ec-fedcf4cecbf1"). InnerVolumeSpecName "kube-api-access-pj4vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.880450 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-util" (OuterVolumeSpecName: "util") pod "e0a75471-aabe-4787-a3ec-fedcf4cecbf1" (UID: "e0a75471-aabe-4787-a3ec-fedcf4cecbf1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.970279 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.970323 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:51 crc kubenswrapper[4696]: I1125 10:50:51.970338 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj4vz\" (UniqueName: \"kubernetes.io/projected/e0a75471-aabe-4787-a3ec-fedcf4cecbf1-kube-api-access-pj4vz\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:52 crc kubenswrapper[4696]: I1125 10:50:52.517268 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" event={"ID":"e0a75471-aabe-4787-a3ec-fedcf4cecbf1","Type":"ContainerDied","Data":"09085930e97cfdc810700157d1aa67ff74d75433952844b07eb29a277c768792"} Nov 25 10:50:52 crc kubenswrapper[4696]: I1125 10:50:52.517314 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09085930e97cfdc810700157d1aa67ff74d75433952844b07eb29a277c768792" Nov 25 10:50:52 crc kubenswrapper[4696]: I1125 10:50:52.517319 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.109930 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn"] Nov 25 10:50:57 crc kubenswrapper[4696]: E1125 10:50:57.110426 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="pull" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.110439 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="pull" Nov 25 10:50:57 crc kubenswrapper[4696]: E1125 10:50:57.110462 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="extract" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.110469 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="extract" Nov 25 10:50:57 crc kubenswrapper[4696]: E1125 10:50:57.110485 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="util" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.110493 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="util" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.110595 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a75471-aabe-4787-a3ec-fedcf4cecbf1" containerName="extract" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.111019 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.117197 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-zz8jr" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.135005 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn"] Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.235892 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nt79\" (UniqueName: \"kubernetes.io/projected/62e71f26-481d-4d1f-b3c7-bed2f91968cd-kube-api-access-6nt79\") pod \"openstack-operator-controller-operator-cccc9d798-j4nfn\" (UID: \"62e71f26-481d-4d1f-b3c7-bed2f91968cd\") " pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.337588 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nt79\" (UniqueName: \"kubernetes.io/projected/62e71f26-481d-4d1f-b3c7-bed2f91968cd-kube-api-access-6nt79\") pod \"openstack-operator-controller-operator-cccc9d798-j4nfn\" (UID: \"62e71f26-481d-4d1f-b3c7-bed2f91968cd\") " pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.361084 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nt79\" (UniqueName: \"kubernetes.io/projected/62e71f26-481d-4d1f-b3c7-bed2f91968cd-kube-api-access-6nt79\") pod \"openstack-operator-controller-operator-cccc9d798-j4nfn\" (UID: \"62e71f26-481d-4d1f-b3c7-bed2f91968cd\") " pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.435089 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:50:57 crc kubenswrapper[4696]: I1125 10:50:57.686906 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn"] Nov 25 10:50:58 crc kubenswrapper[4696]: I1125 10:50:58.564309 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" event={"ID":"62e71f26-481d-4d1f-b3c7-bed2f91968cd","Type":"ContainerStarted","Data":"4fa7b659a52419890ee3bdda56d82cd6adeff8045598850e4f3780eda1662f64"} Nov 25 10:51:00 crc kubenswrapper[4696]: I1125 10:51:00.802211 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:51:00 crc kubenswrapper[4696]: I1125 10:51:00.802547 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:51:00 crc kubenswrapper[4696]: I1125 10:51:00.802616 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:51:00 crc kubenswrapper[4696]: I1125 10:51:00.803392 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7253a2ef071b30ac37f2b8fc9846ceb3e4a68877fb90b28472f1abfa37caf36"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:51:00 crc kubenswrapper[4696]: I1125 10:51:00.803442 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://e7253a2ef071b30ac37f2b8fc9846ceb3e4a68877fb90b28472f1abfa37caf36" gracePeriod=600 Nov 25 10:51:01 crc kubenswrapper[4696]: I1125 10:51:01.586278 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="e7253a2ef071b30ac37f2b8fc9846ceb3e4a68877fb90b28472f1abfa37caf36" exitCode=0 Nov 25 10:51:01 crc kubenswrapper[4696]: I1125 10:51:01.586358 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"e7253a2ef071b30ac37f2b8fc9846ceb3e4a68877fb90b28472f1abfa37caf36"} Nov 25 10:51:01 crc kubenswrapper[4696]: I1125 10:51:01.586609 4696 scope.go:117] "RemoveContainer" containerID="e613ae8723bcb59d28226fb02f66886a567bae947befcc699ab9dd24e0648204" Nov 25 10:51:03 crc kubenswrapper[4696]: I1125 10:51:03.598885 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" event={"ID":"62e71f26-481d-4d1f-b3c7-bed2f91968cd","Type":"ContainerStarted","Data":"62a82bcdfc388eeaf6cae8ddd442de718271cf9e124635bdd930d2baa2e57d7f"} Nov 25 10:51:03 crc kubenswrapper[4696]: I1125 10:51:03.600736 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"33cc4b473fa3af53c05eee091add2db7db56314c26dffae135812139c5ec3acf"} Nov 25 10:51:04 crc kubenswrapper[4696]: I1125 10:51:04.651073 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" podStartSLOduration=1.908184768 podStartE2EDuration="7.651055909s" podCreationTimestamp="2025-11-25 10:50:57 +0000 UTC" firstStartedPulling="2025-11-25 10:50:57.684058078 +0000 UTC m=+854.497674665" lastFinishedPulling="2025-11-25 10:51:03.426929229 +0000 UTC m=+860.240545806" observedRunningTime="2025-11-25 10:51:04.647093564 +0000 UTC m=+861.460710151" watchObservedRunningTime="2025-11-25 10:51:04.651055909 +0000 UTC m=+861.464672496" Nov 25 10:51:05 crc kubenswrapper[4696]: I1125 10:51:05.621187 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:51:17 crc kubenswrapper[4696]: I1125 10:51:17.437315 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-cccc9d798-j4nfn" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.722962 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.725578 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.731149 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6j2tf" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.731398 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.732777 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.740103 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-t2xh6" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.747090 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.767166 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.781791 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.782895 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.786549 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9c5sq" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.810589 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrxzp\" (UniqueName: \"kubernetes.io/projected/c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd-kube-api-access-lrxzp\") pod \"barbican-operator-controller-manager-86dc4d89c8-4kpwn\" (UID: \"c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.810681 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqszw\" (UniqueName: \"kubernetes.io/projected/900bb9ad-0f16-4eb4-beb5-c23e0089db39-kube-api-access-vqszw\") pod \"cinder-operator-controller-manager-79856dc55c-5kdhx\" (UID: \"900bb9ad-0f16-4eb4-beb5-c23e0089db39\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.824877 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.845939 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.846900 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.855445 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-df2dj" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.873183 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.874244 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.877157 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4lf6h" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.895930 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.919213 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-848kc\" (UniqueName: \"kubernetes.io/projected/2ae50512-8525-4878-80b1-7e65a4c035d9-kube-api-access-848kc\") pod \"designate-operator-controller-manager-7d695c9b56-vrf9f\" (UID: \"2ae50512-8525-4878-80b1-7e65a4c035d9\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.919389 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrxzp\" (UniqueName: \"kubernetes.io/projected/c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd-kube-api-access-lrxzp\") pod \"barbican-operator-controller-manager-86dc4d89c8-4kpwn\" (UID: \"c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.919484 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqszw\" (UniqueName: \"kubernetes.io/projected/900bb9ad-0f16-4eb4-beb5-c23e0089db39-kube-api-access-vqszw\") pod \"cinder-operator-controller-manager-79856dc55c-5kdhx\" (UID: \"900bb9ad-0f16-4eb4-beb5-c23e0089db39\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.940184 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c"] Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.970764 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrxzp\" (UniqueName: \"kubernetes.io/projected/c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd-kube-api-access-lrxzp\") pod \"barbican-operator-controller-manager-86dc4d89c8-4kpwn\" (UID: \"c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:51:32 crc kubenswrapper[4696]: I1125 10:51:32.974088 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqszw\" (UniqueName: \"kubernetes.io/projected/900bb9ad-0f16-4eb4-beb5-c23e0089db39-kube-api-access-vqszw\") pod \"cinder-operator-controller-manager-79856dc55c-5kdhx\" (UID: \"900bb9ad-0f16-4eb4-beb5-c23e0089db39\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.008553 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.014469 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.022065 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-klt2c" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.022312 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.023430 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-848kc\" (UniqueName: \"kubernetes.io/projected/2ae50512-8525-4878-80b1-7e65a4c035d9-kube-api-access-848kc\") pod \"designate-operator-controller-manager-7d695c9b56-vrf9f\" (UID: \"2ae50512-8525-4878-80b1-7e65a4c035d9\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.023496 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nxhn\" (UniqueName: \"kubernetes.io/projected/28a8de6d-a7e1-4ca9-853d-6d27a066df12-kube-api-access-6nxhn\") pod \"heat-operator-controller-manager-774b86978c-c7l6c\" (UID: \"28a8de6d-a7e1-4ca9-853d-6d27a066df12\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.023549 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf5f4\" (UniqueName: \"kubernetes.io/projected/7844a3a7-9ef3-48b7-9ab8-944304690567-kube-api-access-qf5f4\") pod \"glance-operator-controller-manager-68b95954c9-ldvdx\" (UID: \"7844a3a7-9ef3-48b7-9ab8-944304690567\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.059798 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.063325 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.063837 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.072117 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-848kc\" (UniqueName: \"kubernetes.io/projected/2ae50512-8525-4878-80b1-7e65a4c035d9-kube-api-access-848kc\") pod \"designate-operator-controller-manager-7d695c9b56-vrf9f\" (UID: \"2ae50512-8525-4878-80b1-7e65a4c035d9\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.094609 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.114078 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.114158 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.114349 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fwv4x" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.125251 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nxhn\" (UniqueName: \"kubernetes.io/projected/28a8de6d-a7e1-4ca9-853d-6d27a066df12-kube-api-access-6nxhn\") pod \"heat-operator-controller-manager-774b86978c-c7l6c\" (UID: \"28a8de6d-a7e1-4ca9-853d-6d27a066df12\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.125293 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpwjz\" (UniqueName: \"kubernetes.io/projected/347bfeb5-b62b-4515-be16-1fdb0fc52e44-kube-api-access-xpwjz\") pod \"horizon-operator-controller-manager-68c9694994-2p89s\" (UID: \"347bfeb5-b62b-4515-be16-1fdb0fc52e44\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.125332 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf5f4\" (UniqueName: \"kubernetes.io/projected/7844a3a7-9ef3-48b7-9ab8-944304690567-kube-api-access-qf5f4\") pod \"glance-operator-controller-manager-68b95954c9-ldvdx\" (UID: \"7844a3a7-9ef3-48b7-9ab8-944304690567\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.164186 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.174686 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.176264 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.188799 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nxhn\" (UniqueName: \"kubernetes.io/projected/28a8de6d-a7e1-4ca9-853d-6d27a066df12-kube-api-access-6nxhn\") pod \"heat-operator-controller-manager-774b86978c-c7l6c\" (UID: \"28a8de6d-a7e1-4ca9-853d-6d27a066df12\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.192787 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.193954 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.194379 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2wchw" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.207500 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf5f4\" (UniqueName: \"kubernetes.io/projected/7844a3a7-9ef3-48b7-9ab8-944304690567-kube-api-access-qf5f4\") pod \"glance-operator-controller-manager-68b95954c9-ldvdx\" (UID: \"7844a3a7-9ef3-48b7-9ab8-944304690567\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.208116 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-wlv86" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.215113 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.226548 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8dc76df-0e30-435e-b61e-926c2b60a87e-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.227324 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpwjz\" (UniqueName: \"kubernetes.io/projected/347bfeb5-b62b-4515-be16-1fdb0fc52e44-kube-api-access-xpwjz\") pod \"horizon-operator-controller-manager-68c9694994-2p89s\" (UID: \"347bfeb5-b62b-4515-be16-1fdb0fc52e44\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.227389 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g286n\" (UniqueName: \"kubernetes.io/projected/f8dc76df-0e30-435e-b61e-926c2b60a87e-kube-api-access-g286n\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.237688 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.244927 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.248035 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.250727 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nldzh" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.252042 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.279300 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpwjz\" (UniqueName: \"kubernetes.io/projected/347bfeb5-b62b-4515-be16-1fdb0fc52e44-kube-api-access-xpwjz\") pod \"horizon-operator-controller-manager-68c9694994-2p89s\" (UID: \"347bfeb5-b62b-4515-be16-1fdb0fc52e44\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.290235 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.291285 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.310752 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.313035 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-g98kk" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.330456 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djdwd\" (UniqueName: \"kubernetes.io/projected/01eb2add-76e2-414f-8a70-275d2a1b8939-kube-api-access-djdwd\") pod \"keystone-operator-controller-manager-748dc6576f-4fkts\" (UID: \"01eb2add-76e2-414f-8a70-275d2a1b8939\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.330515 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8dc76df-0e30-435e-b61e-926c2b60a87e-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.330541 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcwvr\" (UniqueName: \"kubernetes.io/projected/a891b8d2-92ed-4f76-9947-3968ee474921-kube-api-access-qcwvr\") pod \"ironic-operator-controller-manager-5bfcdc958c-q4ml6\" (UID: \"a891b8d2-92ed-4f76-9947-3968ee474921\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.330557 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vbwq\" (UniqueName: \"kubernetes.io/projected/ce501661-3cfd-4cce-8256-c347638da2f6-kube-api-access-9vbwq\") pod \"manila-operator-controller-manager-58bb8d67cc-xwzk7\" (UID: \"ce501661-3cfd-4cce-8256-c347638da2f6\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.330603 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g286n\" (UniqueName: \"kubernetes.io/projected/f8dc76df-0e30-435e-b61e-926c2b60a87e-kube-api-access-g286n\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: E1125 10:51:33.330985 4696 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 10:51:33 crc kubenswrapper[4696]: E1125 10:51:33.331029 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8dc76df-0e30-435e-b61e-926c2b60a87e-cert podName:f8dc76df-0e30-435e-b61e-926c2b60a87e nodeName:}" failed. No retries permitted until 2025-11-25 10:51:33.831014974 +0000 UTC m=+890.644631561 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8dc76df-0e30-435e-b61e-926c2b60a87e-cert") pod "infra-operator-controller-manager-d5cc86f4b-2xtrg" (UID: "f8dc76df-0e30-435e-b61e-926c2b60a87e") : secret "infra-operator-webhook-server-cert" not found Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.352848 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.369730 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.371007 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.381085 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-n2lws" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.381508 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.382472 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.391258 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-csjm5" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.394040 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g286n\" (UniqueName: \"kubernetes.io/projected/f8dc76df-0e30-435e-b61e-926c2b60a87e-kube-api-access-g286n\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.408735 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.409321 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.414978 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.433588 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twsvd\" (UniqueName: \"kubernetes.io/projected/394e9c9d-8d79-4be4-9679-432c658bd52d-kube-api-access-twsvd\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-vdjlx\" (UID: \"394e9c9d-8d79-4be4-9679-432c658bd52d\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.433635 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djdwd\" (UniqueName: \"kubernetes.io/projected/01eb2add-76e2-414f-8a70-275d2a1b8939-kube-api-access-djdwd\") pod \"keystone-operator-controller-manager-748dc6576f-4fkts\" (UID: \"01eb2add-76e2-414f-8a70-275d2a1b8939\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.433697 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcwvr\" (UniqueName: \"kubernetes.io/projected/a891b8d2-92ed-4f76-9947-3968ee474921-kube-api-access-qcwvr\") pod \"ironic-operator-controller-manager-5bfcdc958c-q4ml6\" (UID: \"a891b8d2-92ed-4f76-9947-3968ee474921\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.433716 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vbwq\" (UniqueName: \"kubernetes.io/projected/ce501661-3cfd-4cce-8256-c347638da2f6-kube-api-access-9vbwq\") pod \"manila-operator-controller-manager-58bb8d67cc-xwzk7\" (UID: \"ce501661-3cfd-4cce-8256-c347638da2f6\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.477320 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djdwd\" (UniqueName: \"kubernetes.io/projected/01eb2add-76e2-414f-8a70-275d2a1b8939-kube-api-access-djdwd\") pod \"keystone-operator-controller-manager-748dc6576f-4fkts\" (UID: \"01eb2add-76e2-414f-8a70-275d2a1b8939\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.488684 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.493797 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.495102 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.496297 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vbwq\" (UniqueName: \"kubernetes.io/projected/ce501661-3cfd-4cce-8256-c347638da2f6-kube-api-access-9vbwq\") pod \"manila-operator-controller-manager-58bb8d67cc-xwzk7\" (UID: \"ce501661-3cfd-4cce-8256-c347638da2f6\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.505745 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcwvr\" (UniqueName: \"kubernetes.io/projected/a891b8d2-92ed-4f76-9947-3968ee474921-kube-api-access-qcwvr\") pod \"ironic-operator-controller-manager-5bfcdc958c-q4ml6\" (UID: \"a891b8d2-92ed-4f76-9947-3968ee474921\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.521025 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qxcd9" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.527092 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.536126 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9n9f\" (UniqueName: \"kubernetes.io/projected/8faec67a-3309-496b-9c5a-74eb6421a7f2-kube-api-access-r9n9f\") pod \"neutron-operator-controller-manager-7c57c8bbc4-xqqtv\" (UID: \"8faec67a-3309-496b-9c5a-74eb6421a7f2\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.536221 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twsvd\" (UniqueName: \"kubernetes.io/projected/394e9c9d-8d79-4be4-9679-432c658bd52d-kube-api-access-twsvd\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-vdjlx\" (UID: \"394e9c9d-8d79-4be4-9679-432c658bd52d\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.536275 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt5lf\" (UniqueName: \"kubernetes.io/projected/c40524c0-d668-4b1c-8113-67b152b150e7-kube-api-access-bt5lf\") pod \"nova-operator-controller-manager-79556f57fc-7xgqt\" (UID: \"c40524c0-d668-4b1c-8113-67b152b150e7\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.573059 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.582936 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.583916 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.585418 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twsvd\" (UniqueName: \"kubernetes.io/projected/394e9c9d-8d79-4be4-9679-432c658bd52d-kube-api-access-twsvd\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-vdjlx\" (UID: \"394e9c9d-8d79-4be4-9679-432c658bd52d\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.591044 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.591807 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.592012 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mjd8s" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.595548 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.596750 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.600762 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7td6r" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.619743 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.645105 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.649700 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.656747 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-sjh88" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.660959 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm4ll\" (UniqueName: \"kubernetes.io/projected/c6ca26a1-b284-48b2-b68c-fd0c83b404ec-kube-api-access-hm4ll\") pod \"octavia-operator-controller-manager-fd75fd47d-gjl4t\" (UID: \"c6ca26a1-b284-48b2-b68c-fd0c83b404ec\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.661102 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt5lf\" (UniqueName: \"kubernetes.io/projected/c40524c0-d668-4b1c-8113-67b152b150e7-kube-api-access-bt5lf\") pod \"nova-operator-controller-manager-79556f57fc-7xgqt\" (UID: \"c40524c0-d668-4b1c-8113-67b152b150e7\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.661189 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9n9f\" (UniqueName: \"kubernetes.io/projected/8faec67a-3309-496b-9c5a-74eb6421a7f2-kube-api-access-r9n9f\") pod \"neutron-operator-controller-manager-7c57c8bbc4-xqqtv\" (UID: \"8faec67a-3309-496b-9c5a-74eb6421a7f2\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.665427 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.682980 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.689108 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.771872 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9n9f\" (UniqueName: \"kubernetes.io/projected/8faec67a-3309-496b-9c5a-74eb6421a7f2-kube-api-access-r9n9f\") pod \"neutron-operator-controller-manager-7c57c8bbc4-xqqtv\" (UID: \"8faec67a-3309-496b-9c5a-74eb6421a7f2\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.772331 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt5lf\" (UniqueName: \"kubernetes.io/projected/c40524c0-d668-4b1c-8113-67b152b150e7-kube-api-access-bt5lf\") pod \"nova-operator-controller-manager-79556f57fc-7xgqt\" (UID: \"c40524c0-d668-4b1c-8113-67b152b150e7\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.775339 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm4ll\" (UniqueName: \"kubernetes.io/projected/c6ca26a1-b284-48b2-b68c-fd0c83b404ec-kube-api-access-hm4ll\") pod \"octavia-operator-controller-manager-fd75fd47d-gjl4t\" (UID: \"c6ca26a1-b284-48b2-b68c-fd0c83b404ec\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.775461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n7c9\" (UniqueName: \"kubernetes.io/projected/e6871bcf-31f0-4782-b94c-56e1002cd0d1-kube-api-access-9n7c9\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.775492 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jvr4\" (UniqueName: \"kubernetes.io/projected/7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f-kube-api-access-9jvr4\") pod \"ovn-operator-controller-manager-66cf5c67ff-599kv\" (UID: \"7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.775635 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.775696 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvf6f\" (UniqueName: \"kubernetes.io/projected/33329628-2334-4d45-ad9f-20d2d23ab8ba-kube-api-access-pvf6f\") pod \"placement-operator-controller-manager-5db546f9d9-gnbb8\" (UID: \"33329628-2334-4d45-ad9f-20d2d23ab8ba\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.788941 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.806431 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.824229 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.845152 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.854390 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-72p7l" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.854691 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm4ll\" (UniqueName: \"kubernetes.io/projected/c6ca26a1-b284-48b2-b68c-fd0c83b404ec-kube-api-access-hm4ll\") pod \"octavia-operator-controller-manager-fd75fd47d-gjl4t\" (UID: \"c6ca26a1-b284-48b2-b68c-fd0c83b404ec\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.860154 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.865749 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.887737 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.889068 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.897929 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.902486 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxgfn\" (UniqueName: \"kubernetes.io/projected/33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2-kube-api-access-xxgfn\") pod \"swift-operator-controller-manager-6fdc4fcf86-9ln5q\" (UID: \"33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.902574 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n7c9\" (UniqueName: \"kubernetes.io/projected/e6871bcf-31f0-4782-b94c-56e1002cd0d1-kube-api-access-9n7c9\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.902608 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jvr4\" (UniqueName: \"kubernetes.io/projected/7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f-kube-api-access-9jvr4\") pod \"ovn-operator-controller-manager-66cf5c67ff-599kv\" (UID: \"7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.902648 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8dc76df-0e30-435e-b61e-926c2b60a87e-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.905880 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-tjjlk" Nov 25 10:51:33 crc kubenswrapper[4696]: E1125 10:51:33.905990 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:51:33 crc kubenswrapper[4696]: E1125 10:51:33.906048 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert podName:e6871bcf-31f0-4782-b94c-56e1002cd0d1 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:34.406028548 +0000 UTC m=+891.219645135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" (UID: "e6871bcf-31f0-4782-b94c-56e1002cd0d1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.905894 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.906121 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvf6f\" (UniqueName: \"kubernetes.io/projected/33329628-2334-4d45-ad9f-20d2d23ab8ba-kube-api-access-pvf6f\") pod \"placement-operator-controller-manager-5db546f9d9-gnbb8\" (UID: \"33329628-2334-4d45-ad9f-20d2d23ab8ba\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.906189 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9b5w\" (UniqueName: \"kubernetes.io/projected/90aa0c65-64c1-4ec3-a104-122691247673-kube-api-access-c9b5w\") pod \"telemetry-operator-controller-manager-567f98c9d-jfrk4\" (UID: \"90aa0c65-64c1-4ec3-a104-122691247673\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.913186 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8dc76df-0e30-435e-b61e-926c2b60a87e-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-2xtrg\" (UID: \"f8dc76df-0e30-435e-b61e-926c2b60a87e\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.948584 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4"] Nov 25 10:51:33 crc kubenswrapper[4696]: I1125 10:51:33.987247 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvf6f\" (UniqueName: \"kubernetes.io/projected/33329628-2334-4d45-ad9f-20d2d23ab8ba-kube-api-access-pvf6f\") pod \"placement-operator-controller-manager-5db546f9d9-gnbb8\" (UID: \"33329628-2334-4d45-ad9f-20d2d23ab8ba\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:33.998542 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n7c9\" (UniqueName: \"kubernetes.io/projected/e6871bcf-31f0-4782-b94c-56e1002cd0d1-kube-api-access-9n7c9\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.009656 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxgfn\" (UniqueName: \"kubernetes.io/projected/33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2-kube-api-access-xxgfn\") pod \"swift-operator-controller-manager-6fdc4fcf86-9ln5q\" (UID: \"33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.009833 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9b5w\" (UniqueName: \"kubernetes.io/projected/90aa0c65-64c1-4ec3-a104-122691247673-kube-api-access-c9b5w\") pod \"telemetry-operator-controller-manager-567f98c9d-jfrk4\" (UID: \"90aa0c65-64c1-4ec3-a104-122691247673\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.012978 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.022906 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.035201 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jvr4\" (UniqueName: \"kubernetes.io/projected/7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f-kube-api-access-9jvr4\") pod \"ovn-operator-controller-manager-66cf5c67ff-599kv\" (UID: \"7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.036292 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.047780 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rxkld" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.049911 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9b5w\" (UniqueName: \"kubernetes.io/projected/90aa0c65-64c1-4ec3-a104-122691247673-kube-api-access-c9b5w\") pod \"telemetry-operator-controller-manager-567f98c9d-jfrk4\" (UID: \"90aa0c65-64c1-4ec3-a104-122691247673\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.085313 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxgfn\" (UniqueName: \"kubernetes.io/projected/33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2-kube-api-access-xxgfn\") pod \"swift-operator-controller-manager-6fdc4fcf86-9ln5q\" (UID: \"33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.091944 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.112179 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vtgz\" (UniqueName: \"kubernetes.io/projected/3f468970-3367-47b7-82fa-e9af018c3996-kube-api-access-7vtgz\") pod \"test-operator-controller-manager-5cb74df96-q2c8n\" (UID: \"3f468970-3367-47b7-82fa-e9af018c3996\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.129648 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.133444 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-758cd"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.134696 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.146612 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-82hhk" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.164049 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-758cd"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.207079 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.214720 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r4h4\" (UniqueName: \"kubernetes.io/projected/17643fa7-c097-4a6c-ac8a-fe7ec0cc9901-kube-api-access-4r4h4\") pod \"watcher-operator-controller-manager-864885998-758cd\" (UID: \"17643fa7-c097-4a6c-ac8a-fe7ec0cc9901\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.214768 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vtgz\" (UniqueName: \"kubernetes.io/projected/3f468970-3367-47b7-82fa-e9af018c3996-kube-api-access-7vtgz\") pod \"test-operator-controller-manager-5cb74df96-q2c8n\" (UID: \"3f468970-3367-47b7-82fa-e9af018c3996\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.263510 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vtgz\" (UniqueName: \"kubernetes.io/projected/3f468970-3367-47b7-82fa-e9af018c3996-kube-api-access-7vtgz\") pod \"test-operator-controller-manager-5cb74df96-q2c8n\" (UID: \"3f468970-3367-47b7-82fa-e9af018c3996\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.268440 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.271308 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.277752 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.280068 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.280223 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.280333 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2bw7g" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.284028 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.285308 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.317153 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2ffs\" (UniqueName: \"kubernetes.io/projected/ad6baa72-3698-484a-a2f2-690719d71677-kube-api-access-j2ffs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.317211 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r4h4\" (UniqueName: \"kubernetes.io/projected/17643fa7-c097-4a6c-ac8a-fe7ec0cc9901-kube-api-access-4r4h4\") pod \"watcher-operator-controller-manager-864885998-758cd\" (UID: \"17643fa7-c097-4a6c-ac8a-fe7ec0cc9901\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.317287 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-webhook-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.317328 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.340619 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r4h4\" (UniqueName: \"kubernetes.io/projected/17643fa7-c097-4a6c-ac8a-fe7ec0cc9901-kube-api-access-4r4h4\") pod \"watcher-operator-controller-manager-864885998-758cd\" (UID: \"17643fa7-c097-4a6c-ac8a-fe7ec0cc9901\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.341843 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.342932 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.345772 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-7ksv4" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.354147 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.426555 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2ffs\" (UniqueName: \"kubernetes.io/projected/ad6baa72-3698-484a-a2f2-690719d71677-kube-api-access-j2ffs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.426635 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2s96\" (UniqueName: \"kubernetes.io/projected/3367886a-6027-418f-9415-38d775304079-kube-api-access-m2s96\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94k5m\" (UID: \"3367886a-6027-418f-9415-38d775304079\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.426689 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-webhook-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.426739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.426856 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.426857 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.426914 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs podName:ad6baa72-3698-484a-a2f2-690719d71677 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:34.926893063 +0000 UTC m=+891.740509730 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs") pod "openstack-operator-controller-manager-65d98ff956-rrwtq" (UID: "ad6baa72-3698-484a-a2f2-690719d71677") : secret "metrics-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.427025 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.427067 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert podName:e6871bcf-31f0-4782-b94c-56e1002cd0d1 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:35.427053908 +0000 UTC m=+892.240670495 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" (UID: "e6871bcf-31f0-4782-b94c-56e1002cd0d1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.427164 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.427202 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-webhook-certs podName:ad6baa72-3698-484a-a2f2-690719d71677 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:34.927191932 +0000 UTC m=+891.740808629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-webhook-certs") pod "openstack-operator-controller-manager-65d98ff956-rrwtq" (UID: "ad6baa72-3698-484a-a2f2-690719d71677") : secret "webhook-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.470767 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2ffs\" (UniqueName: \"kubernetes.io/projected/ad6baa72-3698-484a-a2f2-690719d71677-kube-api-access-j2ffs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.530858 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2s96\" (UniqueName: \"kubernetes.io/projected/3367886a-6027-418f-9415-38d775304079-kube-api-access-m2s96\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94k5m\" (UID: \"3367886a-6027-418f-9415-38d775304079\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.548974 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.559252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2s96\" (UniqueName: \"kubernetes.io/projected/3367886a-6027-418f-9415-38d775304079-kube-api-access-m2s96\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94k5m\" (UID: \"3367886a-6027-418f-9415-38d775304079\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.574955 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.668911 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn"] Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.753064 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.865038 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" event={"ID":"c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd","Type":"ContainerStarted","Data":"edc224068519d159989895e87c555e5193fa7bdfe0aa173ca5862aac86856a82"} Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.951495 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-webhook-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.951819 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.952058 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: E1125 10:51:34.952125 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs podName:ad6baa72-3698-484a-a2f2-690719d71677 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:35.952107425 +0000 UTC m=+892.765724012 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs") pod "openstack-operator-controller-manager-65d98ff956-rrwtq" (UID: "ad6baa72-3698-484a-a2f2-690719d71677") : secret "metrics-server-cert" not found Nov 25 10:51:34 crc kubenswrapper[4696]: I1125 10:51:34.981374 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-webhook-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.037564 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.097561 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.101994 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.276843 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.303553 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.340539 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.357373 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.363966 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.497136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.497312 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.497367 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert podName:e6871bcf-31f0-4782-b94c-56e1002cd0d1 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:37.497350419 +0000 UTC m=+894.310967006 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" (UID: "e6871bcf-31f0-4782-b94c-56e1002cd0d1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.700208 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.710927 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.736568 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.751080 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8"] Nov 25 10:51:35 crc kubenswrapper[4696]: W1125 10:51:35.756072 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33329628_2334_4d45_ad9f_20d2d23ab8ba.slice/crio-5878e1bbf0f5372b4490d12da262a584d9e99285bc229c75454582ec7a641e53 WatchSource:0}: Error finding container 5878e1bbf0f5372b4490d12da262a584d9e99285bc229c75454582ec7a641e53: Status 404 returned error can't find the container with id 5878e1bbf0f5372b4490d12da262a584d9e99285bc229c75454582ec7a641e53 Nov 25 10:51:35 crc kubenswrapper[4696]: W1125 10:51:35.759189 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8faec67a_3309_496b_9c5a_74eb6421a7f2.slice/crio-a577e8ef1c19c6ffc37027473888da44725c81ae84c35ad6d38749db45ac5db7 WatchSource:0}: Error finding container a577e8ef1c19c6ffc37027473888da44725c81ae84c35ad6d38749db45ac5db7: Status 404 returned error can't find the container with id a577e8ef1c19c6ffc37027473888da44725c81ae84c35ad6d38749db45ac5db7 Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.772089 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.792366 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx"] Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.795567 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twsvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-cb6c4fdb7-vdjlx_openstack-operators(394e9c9d-8d79-4be4-9679-432c658bd52d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.796764 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xxgfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-9ln5q_openstack-operators(33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.801373 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twsvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-cb6c4fdb7-vdjlx_openstack-operators(394e9c9d-8d79-4be4-9679-432c658bd52d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.801460 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xxgfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-9ln5q_openstack-operators(33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.802782 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" podUID="33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.803366 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" podUID="394e9c9d-8d79-4be4-9679-432c658bd52d" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.834283 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.847095 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.858911 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4"] Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.865035 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4r4h4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-758cd_openstack-operators(17643fa7-c097-4a6c-ac8a-fe7ec0cc9901): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.871594 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-758cd"] Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.871603 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4r4h4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-758cd_openstack-operators(17643fa7-c097-4a6c-ac8a-fe7ec0cc9901): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.872824 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" podUID="17643fa7-c097-4a6c-ac8a-fe7ec0cc9901" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.873440 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jvr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-599kv_openstack-operators(7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: W1125 10:51:35.873592 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f468970_3367_47b7_82fa_e9af018c3996.slice/crio-5f483c20528e447f18cbf084bb8b91f14616f9ad56a7bdc2b1d6e408c7ac2299 WatchSource:0}: Error finding container 5f483c20528e447f18cbf084bb8b91f14616f9ad56a7bdc2b1d6e408c7ac2299: Status 404 returned error can't find the container with id 5f483c20528e447f18cbf084bb8b91f14616f9ad56a7bdc2b1d6e408c7ac2299 Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.875251 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jvr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-599kv_openstack-operators(7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.876791 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" podUID="7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.879691 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7vtgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-q2c8n_openstack-operators(3f468970-3367-47b7-82fa-e9af018c3996): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.890738 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7vtgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-q2c8n_openstack-operators(3f468970-3367-47b7-82fa-e9af018c3996): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.892849 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" podUID="3f468970-3367-47b7-82fa-e9af018c3996" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.893129 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m"] Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.893166 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" event={"ID":"347bfeb5-b62b-4515-be16-1fdb0fc52e44","Type":"ContainerStarted","Data":"7138bde663d5be674af1c26bf7f9fdbd0b807037be8284da77acb57d5973f95d"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.898940 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" event={"ID":"28a8de6d-a7e1-4ca9-853d-6d27a066df12","Type":"ContainerStarted","Data":"4e1861501aee5cfae7e072ce3e1c515133586c444c85b1695850c21c1ea6b2ed"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.902193 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" event={"ID":"17643fa7-c097-4a6c-ac8a-fe7ec0cc9901","Type":"ContainerStarted","Data":"f732f308e064f9112ace44767ba9aae7363ec13babdad75d81497ed5f4e84331"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.904247 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n"] Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.911151 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" podUID="17643fa7-c097-4a6c-ac8a-fe7ec0cc9901" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.911322 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" event={"ID":"a891b8d2-92ed-4f76-9947-3968ee474921","Type":"ContainerStarted","Data":"d33620ec0f80402e1df1f743c4c5d1bab9b75c6df72c5080b6fc707a6877d60d"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.925499 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" event={"ID":"c40524c0-d668-4b1c-8113-67b152b150e7","Type":"ContainerStarted","Data":"7f02d99057c3b9b05fb11be4e4c4f21499bc40d0028be51e53b3f024842bf8f6"} Nov 25 10:51:35 crc kubenswrapper[4696]: W1125 10:51:35.932014 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90aa0c65_64c1_4ec3_a104_122691247673.slice/crio-680724ed989c05f2e818d239285dcaffb6de2adb6f9bb07f1f4f8c663ea173bd WatchSource:0}: Error finding container 680724ed989c05f2e818d239285dcaffb6de2adb6f9bb07f1f4f8c663ea173bd: Status 404 returned error can't find the container with id 680724ed989c05f2e818d239285dcaffb6de2adb6f9bb07f1f4f8c663ea173bd Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.934020 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" event={"ID":"33329628-2334-4d45-ad9f-20d2d23ab8ba","Type":"ContainerStarted","Data":"5878e1bbf0f5372b4490d12da262a584d9e99285bc229c75454582ec7a641e53"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.937357 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" event={"ID":"8faec67a-3309-496b-9c5a-74eb6421a7f2","Type":"ContainerStarted","Data":"a577e8ef1c19c6ffc37027473888da44725c81ae84c35ad6d38749db45ac5db7"} Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.938683 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c9b5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-jfrk4_openstack-operators(90aa0c65-64c1-4ec3-a104-122691247673): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.938818 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" event={"ID":"c6ca26a1-b284-48b2-b68c-fd0c83b404ec","Type":"ContainerStarted","Data":"c423b66df53a477b28470867b6101f86b9343b24b7afae32fab98ea7dd4818a1"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.943392 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" event={"ID":"900bb9ad-0f16-4eb4-beb5-c23e0089db39","Type":"ContainerStarted","Data":"c3083a7eb227d6a70d3bedf637cae3bceae48ee02e610e0f8410f45f514bdfa7"} Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.943463 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c9b5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-jfrk4_openstack-operators(90aa0c65-64c1-4ec3-a104-122691247673): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.944612 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" podUID="90aa0c65-64c1-4ec3-a104-122691247673" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.947329 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" event={"ID":"33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2","Type":"ContainerStarted","Data":"262298c358d57389e6dce6805d1a0f5b4562c343cb8b98822ac6ce95ff62e82e"} Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.949712 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" podUID="33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.951335 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" event={"ID":"7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f","Type":"ContainerStarted","Data":"a2994a50a76b4ff683a26bfd4c3361670cb821246008f087cfe0d317044bd227"} Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.955839 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" podUID="7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f" Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.961014 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" event={"ID":"01eb2add-76e2-414f-8a70-275d2a1b8939","Type":"ContainerStarted","Data":"1c464a4d3e891845a276cc36fb3a7127f484bd76ea5e03583754508e32b391c5"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.963806 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" event={"ID":"3367886a-6027-418f-9415-38d775304079","Type":"ContainerStarted","Data":"3959212405c833473d765fe37699f99cf9acc3a1c6d154ec409e9e1ccceb8389"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.972655 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" event={"ID":"7844a3a7-9ef3-48b7-9ab8-944304690567","Type":"ContainerStarted","Data":"1404e70b517e63886a14b775091b19bb3ec2b7fe3ba8b2c0e582f139ec19a5da"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.978278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" event={"ID":"ce501661-3cfd-4cce-8256-c347638da2f6","Type":"ContainerStarted","Data":"7b1d18f6d8ee02ea62aa6f8f45c757cbed7f25a0705fa8d236f616098af58a67"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.985110 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" event={"ID":"2ae50512-8525-4878-80b1-7e65a4c035d9","Type":"ContainerStarted","Data":"6dadadaa5f76040bb73247cec47fe129c12e59184b2237f8b71b32c9d33d6830"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.986787 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" event={"ID":"394e9c9d-8d79-4be4-9679-432c658bd52d","Type":"ContainerStarted","Data":"f4088190c84034262a842a71ca2d4e6b09c380488e075298a79e759f9df9c836"} Nov 25 10:51:35 crc kubenswrapper[4696]: I1125 10:51:35.988965 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" event={"ID":"f8dc76df-0e30-435e-b61e-926c2b60a87e","Type":"ContainerStarted","Data":"f29b196cb81c4b2d7008631dba71aa4d153c0f90587b07cbe185050e4bc092d3"} Nov 25 10:51:35 crc kubenswrapper[4696]: E1125 10:51:35.991349 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" podUID="394e9c9d-8d79-4be4-9679-432c658bd52d" Nov 25 10:51:36 crc kubenswrapper[4696]: I1125 10:51:36.014461 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:36 crc kubenswrapper[4696]: E1125 10:51:36.015475 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:51:36 crc kubenswrapper[4696]: E1125 10:51:36.015529 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs podName:ad6baa72-3698-484a-a2f2-690719d71677 nodeName:}" failed. No retries permitted until 2025-11-25 10:51:38.015512438 +0000 UTC m=+894.829129025 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs") pod "openstack-operator-controller-manager-65d98ff956-rrwtq" (UID: "ad6baa72-3698-484a-a2f2-690719d71677") : secret "metrics-server-cert" not found Nov 25 10:51:37 crc kubenswrapper[4696]: I1125 10:51:37.035203 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" event={"ID":"3f468970-3367-47b7-82fa-e9af018c3996","Type":"ContainerStarted","Data":"5f483c20528e447f18cbf084bb8b91f14616f9ad56a7bdc2b1d6e408c7ac2299"} Nov 25 10:51:37 crc kubenswrapper[4696]: I1125 10:51:37.051720 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" event={"ID":"90aa0c65-64c1-4ec3-a104-122691247673","Type":"ContainerStarted","Data":"680724ed989c05f2e818d239285dcaffb6de2adb6f9bb07f1f4f8c663ea173bd"} Nov 25 10:51:37 crc kubenswrapper[4696]: E1125 10:51:37.058792 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" podUID="33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2" Nov 25 10:51:37 crc kubenswrapper[4696]: E1125 10:51:37.065706 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" podUID="17643fa7-c097-4a6c-ac8a-fe7ec0cc9901" Nov 25 10:51:37 crc kubenswrapper[4696]: E1125 10:51:37.065789 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" podUID="90aa0c65-64c1-4ec3-a104-122691247673" Nov 25 10:51:37 crc kubenswrapper[4696]: E1125 10:51:37.065841 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" podUID="394e9c9d-8d79-4be4-9679-432c658bd52d" Nov 25 10:51:37 crc kubenswrapper[4696]: E1125 10:51:37.065881 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" podUID="7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f" Nov 25 10:51:37 crc kubenswrapper[4696]: E1125 10:51:37.098880 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" podUID="3f468970-3367-47b7-82fa-e9af018c3996" Nov 25 10:51:37 crc kubenswrapper[4696]: I1125 10:51:37.543181 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:37 crc kubenswrapper[4696]: I1125 10:51:37.562465 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6871bcf-31f0-4782-b94c-56e1002cd0d1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t\" (UID: \"e6871bcf-31f0-4782-b94c-56e1002cd0d1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:37 crc kubenswrapper[4696]: I1125 10:51:37.846627 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:51:38 crc kubenswrapper[4696]: I1125 10:51:38.051991 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:38 crc kubenswrapper[4696]: I1125 10:51:38.056547 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad6baa72-3698-484a-a2f2-690719d71677-metrics-certs\") pod \"openstack-operator-controller-manager-65d98ff956-rrwtq\" (UID: \"ad6baa72-3698-484a-a2f2-690719d71677\") " pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:38 crc kubenswrapper[4696]: E1125 10:51:38.057789 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" podUID="90aa0c65-64c1-4ec3-a104-122691247673" Nov 25 10:51:38 crc kubenswrapper[4696]: E1125 10:51:38.058755 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" podUID="3f468970-3367-47b7-82fa-e9af018c3996" Nov 25 10:51:38 crc kubenswrapper[4696]: I1125 10:51:38.317317 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:51:48 crc kubenswrapper[4696]: I1125 10:51:48.043656 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:51:54 crc kubenswrapper[4696]: E1125 10:51:54.837634 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c" Nov 25 10:51:54 crc kubenswrapper[4696]: E1125 10:51:54.839357 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pvf6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5db546f9d9-gnbb8_openstack-operators(33329628-2334-4d45-ad9f-20d2d23ab8ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:51:55 crc kubenswrapper[4696]: E1125 10:51:55.713773 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9" Nov 25 10:51:55 crc kubenswrapper[4696]: E1125 10:51:55.714203 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vqszw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-79856dc55c-5kdhx_openstack-operators(900bb9ad-0f16-4eb4-beb5-c23e0089db39): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:51:58 crc kubenswrapper[4696]: E1125 10:51:58.520235 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6" Nov 25 10:51:58 crc kubenswrapper[4696]: E1125 10:51:58.520638 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9n9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7c57c8bbc4-xqqtv_openstack-operators(8faec67a-3309-496b-9c5a-74eb6421a7f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:51:59 crc kubenswrapper[4696]: E1125 10:51:59.256934 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894" Nov 25 10:51:59 crc kubenswrapper[4696]: E1125 10:51:59.257084 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g286n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-d5cc86f4b-2xtrg_openstack-operators(f8dc76df-0e30-435e-b61e-926c2b60a87e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:51:59 crc kubenswrapper[4696]: E1125 10:51:59.575786 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9" Nov 25 10:51:59 crc kubenswrapper[4696]: E1125 10:51:59.576431 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xpwjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c9694994-2p89s_openstack-operators(347bfeb5-b62b-4515-be16-1fdb0fc52e44): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:52:00 crc kubenswrapper[4696]: E1125 10:52:00.366520 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13" Nov 25 10:52:00 crc kubenswrapper[4696]: E1125 10:52:00.366744 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hm4ll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-gjl4t_openstack-operators(c6ca26a1-b284-48b2-b68c-fd0c83b404ec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:52:07 crc kubenswrapper[4696]: I1125 10:52:07.199008 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t"] Nov 25 10:52:09 crc kubenswrapper[4696]: E1125 10:52:09.428009 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 25 10:52:09 crc kubenswrapper[4696]: E1125 10:52:09.428439 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2s96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-94k5m_openstack-operators(3367886a-6027-418f-9415-38d775304079): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:52:09 crc kubenswrapper[4696]: E1125 10:52:09.429874 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" podUID="3367886a-6027-418f-9415-38d775304079" Nov 25 10:52:10 crc kubenswrapper[4696]: I1125 10:52:10.332948 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq"] Nov 25 10:52:10 crc kubenswrapper[4696]: I1125 10:52:10.358124 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" event={"ID":"c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd","Type":"ContainerStarted","Data":"bc40ab1a342781f986e4e35e91e2014afb3ff5c83f1d7c8aa5e5304ebf27eb4b"} Nov 25 10:52:10 crc kubenswrapper[4696]: I1125 10:52:10.363022 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" event={"ID":"e6871bcf-31f0-4782-b94c-56e1002cd0d1","Type":"ContainerStarted","Data":"e2c3770806fa382fdba2ffcb9f0e311130c74be5e1330c58364772d57c308a10"} Nov 25 10:52:10 crc kubenswrapper[4696]: E1125 10:52:10.368140 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" podUID="3367886a-6027-418f-9415-38d775304079" Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.394317 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" event={"ID":"ce501661-3cfd-4cce-8256-c347638da2f6","Type":"ContainerStarted","Data":"6ec7717bf7a7542728f5e98c39b0e0a02f6f025a0684dfdd7dfdd6ca445f0aa7"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.395348 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" event={"ID":"c40524c0-d668-4b1c-8113-67b152b150e7","Type":"ContainerStarted","Data":"93193efe0b2e69b58c476b1b0b4519f71cb84e3e7af3a5c933ea45f7f62d93db"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.408447 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" event={"ID":"90aa0c65-64c1-4ec3-a104-122691247673","Type":"ContainerStarted","Data":"734ba55a5376d1e9d0763bfdbfdcba93484159e4f5e239ad5fcb780bbc324d6d"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.416250 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" event={"ID":"2ae50512-8525-4878-80b1-7e65a4c035d9","Type":"ContainerStarted","Data":"c2af6bb3993746e6d23f50dea88f08fd9db3a5ccc951b1ac767962f08cc2acb7"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.426367 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" event={"ID":"28a8de6d-a7e1-4ca9-853d-6d27a066df12","Type":"ContainerStarted","Data":"772b5aefed1c025b89c510c3ba1af6825c86470af905a04c224358c9c4cd8444"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.448125 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" event={"ID":"7844a3a7-9ef3-48b7-9ab8-944304690567","Type":"ContainerStarted","Data":"54273b44274b06757e0ef8d270e2a552361a8a2d8700880068ed15e245f4d100"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.449430 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" event={"ID":"a891b8d2-92ed-4f76-9947-3968ee474921","Type":"ContainerStarted","Data":"03d7c443bc74c0f3639a43aefc05261f0de5f8a81432528774c46c14da5dc3e6"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.450687 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" event={"ID":"ad6baa72-3698-484a-a2f2-690719d71677","Type":"ContainerStarted","Data":"811278d0a9c8fc4f83c902f16837ad3a6a2bb590dc4995449e770d6bbc498eeb"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.451922 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" event={"ID":"3f468970-3367-47b7-82fa-e9af018c3996","Type":"ContainerStarted","Data":"04094f57c5d435588dd45ed22ab346b9d22e6d7ad20f15bb127dadc5cab3bf9e"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.453118 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" event={"ID":"33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2","Type":"ContainerStarted","Data":"7106d55875785d4ab411e2d217b34ec142b14943eb2660be532836c2f772c672"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.466895 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" event={"ID":"7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f","Type":"ContainerStarted","Data":"5068f9f62519731bc1640a5b40eb60c70f0fba96c40fa607afa476bbf72267cf"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.471154 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" event={"ID":"394e9c9d-8d79-4be4-9679-432c658bd52d","Type":"ContainerStarted","Data":"528586580c28a1ec8fc10616321b91e2e07c89b7806e4cdd1579918afc9ec4b9"} Nov 25 10:52:11 crc kubenswrapper[4696]: I1125 10:52:11.489894 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" event={"ID":"01eb2add-76e2-414f-8a70-275d2a1b8939","Type":"ContainerStarted","Data":"4c364403e524c7668f426d5cffc7de373bd4c6154fed6ebc308620fd057821cc"} Nov 25 10:52:13 crc kubenswrapper[4696]: I1125 10:52:13.504159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" event={"ID":"17643fa7-c097-4a6c-ac8a-fe7ec0cc9901","Type":"ContainerStarted","Data":"daaeb3aec8730047e7ee5d7401b306e729caf852096d46c65cdc0eb18e6aa6c2"} Nov 25 10:52:13 crc kubenswrapper[4696]: I1125 10:52:13.506163 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" event={"ID":"ad6baa72-3698-484a-a2f2-690719d71677","Type":"ContainerStarted","Data":"76585e240623e1b7c479eab9a090c75e67dd13f4b3a1747b8b162b2444beaecb"} Nov 25 10:52:13 crc kubenswrapper[4696]: I1125 10:52:13.506325 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:52:13 crc kubenswrapper[4696]: I1125 10:52:13.546590 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" podStartSLOduration=40.546574143 podStartE2EDuration="40.546574143s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:13.544456377 +0000 UTC m=+930.358072984" watchObservedRunningTime="2025-11-25 10:52:13.546574143 +0000 UTC m=+930.360190720" Nov 25 10:52:16 crc kubenswrapper[4696]: E1125 10:52:16.376037 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" podUID="347bfeb5-b62b-4515-be16-1fdb0fc52e44" Nov 25 10:52:16 crc kubenswrapper[4696]: I1125 10:52:16.530121 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" event={"ID":"347bfeb5-b62b-4515-be16-1fdb0fc52e44","Type":"ContainerStarted","Data":"3b298bb16be215436b9887aca83fd0e0f49602cb289262695238ee83aaccd30c"} Nov 25 10:52:16 crc kubenswrapper[4696]: I1125 10:52:16.538892 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" event={"ID":"e6871bcf-31f0-4782-b94c-56e1002cd0d1","Type":"ContainerStarted","Data":"da14e2eb44e2fe8e1e6dbcf697378a1784fea5149ca587bbf979d93301872cb7"} Nov 25 10:52:16 crc kubenswrapper[4696]: E1125 10:52:16.584380 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" podUID="f8dc76df-0e30-435e-b61e-926c2b60a87e" Nov 25 10:52:16 crc kubenswrapper[4696]: E1125 10:52:16.744890 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" podUID="33329628-2334-4d45-ad9f-20d2d23ab8ba" Nov 25 10:52:16 crc kubenswrapper[4696]: E1125 10:52:16.745967 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" podUID="900bb9ad-0f16-4eb4-beb5-c23e0089db39" Nov 25 10:52:17 crc kubenswrapper[4696]: E1125 10:52:17.266648 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" podUID="c6ca26a1-b284-48b2-b68c-fd0c83b404ec" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.563610 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" event={"ID":"7844a3a7-9ef3-48b7-9ab8-944304690567","Type":"ContainerStarted","Data":"2d47141eab6721b885e5ee05fad56df4a7e2153313bc138a5f08c34c75199d22"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.565376 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.567803 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.587010 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" event={"ID":"ce501661-3cfd-4cce-8256-c347638da2f6","Type":"ContainerStarted","Data":"1e07da77c9c6c7bcbe7e5d20465a82d89e56591b4479b36cc6d9906b2c869f11"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.587115 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.592610 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" event={"ID":"90aa0c65-64c1-4ec3-a104-122691247673","Type":"ContainerStarted","Data":"bedc477b305c54f193f5f7c8fd763fe0f6c88c9a3b3cb9a2a0fb85c28ebd2d3b"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.592950 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.594090 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.599061 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" event={"ID":"33329628-2334-4d45-ad9f-20d2d23ab8ba","Type":"ContainerStarted","Data":"db94850dea7730ed1e09c8ad2a4c32f9408e5d3cfca47207b77fd9fbb2fd68b8"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.599558 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ldvdx" podStartSLOduration=4.752691819 podStartE2EDuration="45.599547346s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.306769129 +0000 UTC m=+892.120385716" lastFinishedPulling="2025-11-25 10:52:16.153624656 +0000 UTC m=+932.967241243" observedRunningTime="2025-11-25 10:52:17.594210567 +0000 UTC m=+934.407827164" watchObservedRunningTime="2025-11-25 10:52:17.599547346 +0000 UTC m=+934.413163933" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.605148 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.615442 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" event={"ID":"2ae50512-8525-4878-80b1-7e65a4c035d9","Type":"ContainerStarted","Data":"95bd05115bc3c088dc0a61de101cb6148ca1c26837d9a62300447292391b640c"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.616287 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.650910 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.658205 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" event={"ID":"394e9c9d-8d79-4be4-9679-432c658bd52d","Type":"ContainerStarted","Data":"9f8248a90e6af4f5bf0191e19073f2f0c47f32862a2133941b22f46ead2e16ce"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.659155 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.662616 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-jfrk4" podStartSLOduration=4.321059132 podStartE2EDuration="44.662597061s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.938565377 +0000 UTC m=+892.752181964" lastFinishedPulling="2025-11-25 10:52:16.280103296 +0000 UTC m=+933.093719893" observedRunningTime="2025-11-25 10:52:17.650695754 +0000 UTC m=+934.464312341" watchObservedRunningTime="2025-11-25 10:52:17.662597061 +0000 UTC m=+934.476213658" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.670630 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" event={"ID":"28a8de6d-a7e1-4ca9-853d-6d27a066df12","Type":"ContainerStarted","Data":"6f94abcca7a6d9438113d4083a545d5f3be452f92fb7a27d6245b807a42e64f1"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.671840 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.673983 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.674271 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.689491 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-xwzk7" podStartSLOduration=3.179714232 podStartE2EDuration="44.689470271s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.361852559 +0000 UTC m=+892.175469146" lastFinishedPulling="2025-11-25 10:52:16.871608588 +0000 UTC m=+933.685225185" observedRunningTime="2025-11-25 10:52:17.686172946 +0000 UTC m=+934.499789533" watchObservedRunningTime="2025-11-25 10:52:17.689470271 +0000 UTC m=+934.503086858" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.715614 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" event={"ID":"a891b8d2-92ed-4f76-9947-3968ee474921","Type":"ContainerStarted","Data":"b9176dfcc00d30e95ac612e9f52482133e82a838d43a33765d0e8ea9631f541c"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.716862 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.724749 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-c7l6c" podStartSLOduration=4.596637974 podStartE2EDuration="45.724734236s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.056302467 +0000 UTC m=+891.869919054" lastFinishedPulling="2025-11-25 10:52:16.184398739 +0000 UTC m=+932.998015316" observedRunningTime="2025-11-25 10:52:17.724041005 +0000 UTC m=+934.537657592" watchObservedRunningTime="2025-11-25 10:52:17.724734236 +0000 UTC m=+934.538350823" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.731981 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" event={"ID":"900bb9ad-0f16-4eb4-beb5-c23e0089db39","Type":"ContainerStarted","Data":"cf2bc1b9fe02369ede3f6bc0afd58e7ea53203a9a4f0cc8fa80765cfb57d6ee7"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.750506 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.756583 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" event={"ID":"c6ca26a1-b284-48b2-b68c-fd0c83b404ec","Type":"ContainerStarted","Data":"bef6ed6b98c01af9fb6fb353fcb9d2af46620c9bcfdefb191ede4f150de8d9f3"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.780594 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" event={"ID":"f8dc76df-0e30-435e-b61e-926c2b60a87e","Type":"ContainerStarted","Data":"3ad1e85c673359ad47e38433c621efa63421a3cfd2a0dea2bbba7cb6463a6bf3"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.813842 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.814764 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" event={"ID":"e6871bcf-31f0-4782-b94c-56e1002cd0d1","Type":"ContainerStarted","Data":"86ec3388b011a2fd50d9422de2d0aae3f8b6bb045da2d0715af1a103bb83aa86"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.832446 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-vdjlx" podStartSLOduration=4.472061245 podStartE2EDuration="44.832429093s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.795358154 +0000 UTC m=+892.608974741" lastFinishedPulling="2025-11-25 10:52:16.155726002 +0000 UTC m=+932.969342589" observedRunningTime="2025-11-25 10:52:17.830572215 +0000 UTC m=+934.644188812" watchObservedRunningTime="2025-11-25 10:52:17.832429093 +0000 UTC m=+934.646045680" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.850892 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" event={"ID":"3f468970-3367-47b7-82fa-e9af018c3996","Type":"ContainerStarted","Data":"cbef941d7e31a0c27d6ba5e0775ec2cb8a9b628274df2805549566ee66771863"} Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.851896 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.858913 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.890289 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-vrf9f" podStartSLOduration=3.922684742 podStartE2EDuration="45.890268023s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.10453611 +0000 UTC m=+891.918152697" lastFinishedPulling="2025-11-25 10:52:17.072119391 +0000 UTC m=+933.885735978" observedRunningTime="2025-11-25 10:52:17.88829146 +0000 UTC m=+934.701908047" watchObservedRunningTime="2025-11-25 10:52:17.890268023 +0000 UTC m=+934.703884610" Nov 25 10:52:17 crc kubenswrapper[4696]: I1125 10:52:17.942076 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" podStartSLOduration=38.382456836 podStartE2EDuration="44.942060241s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:52:09.40687651 +0000 UTC m=+926.220493097" lastFinishedPulling="2025-11-25 10:52:15.966479915 +0000 UTC m=+932.780096502" observedRunningTime="2025-11-25 10:52:17.93411389 +0000 UTC m=+934.747730497" watchObservedRunningTime="2025-11-25 10:52:17.942060241 +0000 UTC m=+934.755676828" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.032762 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-q2c8n" podStartSLOduration=4.316090457 podStartE2EDuration="45.032736179s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.879589745 +0000 UTC m=+892.693206332" lastFinishedPulling="2025-11-25 10:52:16.596235467 +0000 UTC m=+933.409852054" observedRunningTime="2025-11-25 10:52:18.020895786 +0000 UTC m=+934.834512393" watchObservedRunningTime="2025-11-25 10:52:18.032736179 +0000 UTC m=+934.846352766" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.152575 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-q4ml6" podStartSLOduration=4.419940203 podStartE2EDuration="46.152560401s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.340638619 +0000 UTC m=+892.154255206" lastFinishedPulling="2025-11-25 10:52:17.073258827 +0000 UTC m=+933.886875404" observedRunningTime="2025-11-25 10:52:18.09848946 +0000 UTC m=+934.912106057" watchObservedRunningTime="2025-11-25 10:52:18.152560401 +0000 UTC m=+934.966176988" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.327687 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-65d98ff956-rrwtq" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.859387 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" event={"ID":"347bfeb5-b62b-4515-be16-1fdb0fc52e44","Type":"ContainerStarted","Data":"1f66e38630ad1af739033ca524526f1f72e789fe073d79d900f7924f94f87ed8"} Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.860312 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.861773 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" event={"ID":"c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd","Type":"ContainerStarted","Data":"0cbf986d09b3d41fc92b895f5f9c2eb4d7e638f6b5c5bca0ed08ca96dc894271"} Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.862310 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.864266 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" event={"ID":"7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f","Type":"ContainerStarted","Data":"98a2da4ef9572318f34fa9976d84b91979b7a5390398cdcc38a81000f496ebc3"} Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.864771 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.865417 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.866464 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" event={"ID":"17643fa7-c097-4a6c-ac8a-fe7ec0cc9901","Type":"ContainerStarted","Data":"98c9e3a4ae38016773410750fe02c352bae7a02d8eba5df254e5477dbe5fa4b1"} Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.866933 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.868311 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" event={"ID":"c40524c0-d668-4b1c-8113-67b152b150e7","Type":"ContainerStarted","Data":"797f1bc952f09c7dba738120f063674377aff24dee20d6a036a261af6064d05b"} Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.868769 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.870904 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" event={"ID":"33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2","Type":"ContainerStarted","Data":"35922f960467cb3aea4e4327e1f7b6e984190f81cf3025f0f12113bd43407881"} Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.871191 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.872678 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.873727 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.873878 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.880841 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.887135 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" podStartSLOduration=4.941616534 podStartE2EDuration="46.887116437s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.364606346 +0000 UTC m=+892.178222933" lastFinishedPulling="2025-11-25 10:52:17.310106249 +0000 UTC m=+934.123722836" observedRunningTime="2025-11-25 10:52:18.885385003 +0000 UTC m=+935.699001600" watchObservedRunningTime="2025-11-25 10:52:18.887116437 +0000 UTC m=+935.700733034" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.919191 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-7xgqt" podStartSLOduration=3.712430388 podStartE2EDuration="45.919166202s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.723014788 +0000 UTC m=+892.536631375" lastFinishedPulling="2025-11-25 10:52:17.929750612 +0000 UTC m=+934.743367189" observedRunningTime="2025-11-25 10:52:18.914155262 +0000 UTC m=+935.727771859" watchObservedRunningTime="2025-11-25 10:52:18.919166202 +0000 UTC m=+935.732782789" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.957488 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-9ln5q" podStartSLOduration=4.394805121 podStartE2EDuration="45.957465023s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.796627134 +0000 UTC m=+892.610243721" lastFinishedPulling="2025-11-25 10:52:17.359287036 +0000 UTC m=+934.172903623" observedRunningTime="2025-11-25 10:52:18.944258645 +0000 UTC m=+935.757875242" watchObservedRunningTime="2025-11-25 10:52:18.957465023 +0000 UTC m=+935.771081610" Nov 25 10:52:18 crc kubenswrapper[4696]: I1125 10:52:18.994434 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-599kv" podStartSLOduration=4.086296679 podStartE2EDuration="45.994411332s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.873250965 +0000 UTC m=+892.686867562" lastFinishedPulling="2025-11-25 10:52:17.781365628 +0000 UTC m=+934.594982215" observedRunningTime="2025-11-25 10:52:18.989460585 +0000 UTC m=+935.803077182" watchObservedRunningTime="2025-11-25 10:52:18.994411332 +0000 UTC m=+935.808027919" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.037982 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-4kpwn" podStartSLOduration=4.4670902 podStartE2EDuration="47.037707461s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:34.72182954 +0000 UTC m=+891.535446137" lastFinishedPulling="2025-11-25 10:52:17.292446811 +0000 UTC m=+934.106063398" observedRunningTime="2025-11-25 10:52:19.030249725 +0000 UTC m=+935.843866332" watchObservedRunningTime="2025-11-25 10:52:19.037707461 +0000 UTC m=+935.851324068" Nov 25 10:52:19 crc kubenswrapper[4696]: E1125 10:52:19.267975 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" podUID="8faec67a-3309-496b-9c5a-74eb6421a7f2" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.879340 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" event={"ID":"8faec67a-3309-496b-9c5a-74eb6421a7f2","Type":"ContainerStarted","Data":"9b2dfbca5a03ee294385338ab681dcf2f8f999520949a1f2959779c41caa1ebb"} Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.884941 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" event={"ID":"c6ca26a1-b284-48b2-b68c-fd0c83b404ec","Type":"ContainerStarted","Data":"e5c97ff5c9246e8a2770e29bb9e76d855d1a62708b950a234fe1ebcf99a84a73"} Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.885088 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.888025 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" event={"ID":"01eb2add-76e2-414f-8a70-275d2a1b8939","Type":"ContainerStarted","Data":"e3254168c953d43a764f0bed5c53cac87416fc4530363cc363755edda830841d"} Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.888196 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.890044 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" event={"ID":"f8dc76df-0e30-435e-b61e-926c2b60a87e","Type":"ContainerStarted","Data":"0db62476e84c45a48b464ca7f9b656bce24d61ec56045e1fa6101261291df260"} Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.890178 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.890799 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.891988 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" event={"ID":"33329628-2334-4d45-ad9f-20d2d23ab8ba","Type":"ContainerStarted","Data":"a0d4266809ac8ed242fa4e0c05366ccee150b0e56a0f52933f729e97480723b7"} Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.892091 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.894430 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" event={"ID":"900bb9ad-0f16-4eb4-beb5-c23e0089db39","Type":"ContainerStarted","Data":"9d33791c93686fc4457e045e2c75c06417f92e9ed01fc73cb9ad6972dc3838fd"} Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.894457 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.902522 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-758cd" podStartSLOduration=5.220119686 podStartE2EDuration="46.902504788s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.864912411 +0000 UTC m=+892.678528998" lastFinishedPulling="2025-11-25 10:52:17.547297523 +0000 UTC m=+934.360914100" observedRunningTime="2025-11-25 10:52:19.089844291 +0000 UTC m=+935.903460878" watchObservedRunningTime="2025-11-25 10:52:19.902504788 +0000 UTC m=+936.716121375" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.928158 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-4fkts" podStartSLOduration=4.415839093 podStartE2EDuration="47.928137919s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.30712748 +0000 UTC m=+892.120744067" lastFinishedPulling="2025-11-25 10:52:18.819426306 +0000 UTC m=+935.633042893" observedRunningTime="2025-11-25 10:52:19.919403713 +0000 UTC m=+936.733020310" watchObservedRunningTime="2025-11-25 10:52:19.928137919 +0000 UTC m=+936.741754506" Nov 25 10:52:19 crc kubenswrapper[4696]: I1125 10:52:19.987312 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" podStartSLOduration=3.267677756 podStartE2EDuration="46.987295061s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.764891881 +0000 UTC m=+892.578508468" lastFinishedPulling="2025-11-25 10:52:19.484509176 +0000 UTC m=+936.298125773" observedRunningTime="2025-11-25 10:52:19.953113829 +0000 UTC m=+936.766730416" watchObservedRunningTime="2025-11-25 10:52:19.987295061 +0000 UTC m=+936.800911648" Nov 25 10:52:20 crc kubenswrapper[4696]: I1125 10:52:20.012880 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" podStartSLOduration=4.291703107 podStartE2EDuration="48.012858059s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.769817197 +0000 UTC m=+892.583433784" lastFinishedPulling="2025-11-25 10:52:19.490972149 +0000 UTC m=+936.304588736" observedRunningTime="2025-11-25 10:52:19.988285022 +0000 UTC m=+936.801901619" watchObservedRunningTime="2025-11-25 10:52:20.012858059 +0000 UTC m=+936.826474646" Nov 25 10:52:20 crc kubenswrapper[4696]: I1125 10:52:20.074617 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" podStartSLOduration=3.613997928 podStartE2EDuration="48.074595722s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.059643682 +0000 UTC m=+891.873260269" lastFinishedPulling="2025-11-25 10:52:19.520241476 +0000 UTC m=+936.333858063" observedRunningTime="2025-11-25 10:52:20.069539883 +0000 UTC m=+936.883156460" watchObservedRunningTime="2025-11-25 10:52:20.074595722 +0000 UTC m=+936.888212319" Nov 25 10:52:20 crc kubenswrapper[4696]: I1125 10:52:20.076713 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" podStartSLOduration=3.313969642 podStartE2EDuration="47.076697909s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.756136595 +0000 UTC m=+892.569753182" lastFinishedPulling="2025-11-25 10:52:19.518864862 +0000 UTC m=+936.332481449" observedRunningTime="2025-11-25 10:52:20.016280938 +0000 UTC m=+936.829897535" watchObservedRunningTime="2025-11-25 10:52:20.076697909 +0000 UTC m=+936.890314516" Nov 25 10:52:20 crc kubenswrapper[4696]: I1125 10:52:20.902205 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" event={"ID":"8faec67a-3309-496b-9c5a-74eb6421a7f2","Type":"ContainerStarted","Data":"f59903335fd742d66ab7fa3bb23d840130737388fc4670d6b3354fcff93941fd"} Nov 25 10:52:20 crc kubenswrapper[4696]: I1125 10:52:20.922573 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" podStartSLOduration=3.3265231379999998 podStartE2EDuration="47.922555597s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.769933231 +0000 UTC m=+892.583549818" lastFinishedPulling="2025-11-25 10:52:20.36596569 +0000 UTC m=+937.179582277" observedRunningTime="2025-11-25 10:52:20.918344764 +0000 UTC m=+937.731961361" watchObservedRunningTime="2025-11-25 10:52:20.922555597 +0000 UTC m=+937.736172184" Nov 25 10:52:21 crc kubenswrapper[4696]: I1125 10:52:21.908036 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:52:23 crc kubenswrapper[4696]: I1125 10:52:23.413050 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2p89s" Nov 25 10:52:24 crc kubenswrapper[4696]: I1125 10:52:24.026613 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-gnbb8" Nov 25 10:52:24 crc kubenswrapper[4696]: I1125 10:52:24.100007 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-2xtrg" Nov 25 10:52:25 crc kubenswrapper[4696]: I1125 10:52:25.934198 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" event={"ID":"3367886a-6027-418f-9415-38d775304079","Type":"ContainerStarted","Data":"a93d40e4f6afe3f3e02ffa65f0b19f444a739e3dcd0cac83c8e768331637e32b"} Nov 25 10:52:25 crc kubenswrapper[4696]: I1125 10:52:25.951393 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94k5m" podStartSLOduration=2.34061541 podStartE2EDuration="51.951375491s" podCreationTimestamp="2025-11-25 10:51:34 +0000 UTC" firstStartedPulling="2025-11-25 10:51:35.864483317 +0000 UTC m=+892.678099904" lastFinishedPulling="2025-11-25 10:52:25.475243398 +0000 UTC m=+942.288859985" observedRunningTime="2025-11-25 10:52:25.948583583 +0000 UTC m=+942.762200170" watchObservedRunningTime="2025-11-25 10:52:25.951375491 +0000 UTC m=+942.764992078" Nov 25 10:52:27 crc kubenswrapper[4696]: I1125 10:52:27.854456 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t" Nov 25 10:52:33 crc kubenswrapper[4696]: I1125 10:52:33.066970 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-5kdhx" Nov 25 10:52:33 crc kubenswrapper[4696]: I1125 10:52:33.793713 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-xqqtv" Nov 25 10:52:33 crc kubenswrapper[4696]: I1125 10:52:33.864748 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-gjl4t" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.182098 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-945jr"] Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.184241 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.186812 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.187218 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.187262 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-zx47h" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.192404 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.205117 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-945jr"] Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.293001 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d5phf"] Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.294082 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.298335 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.321636 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d5phf"] Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.355010 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63236751-76a1-457c-9885-6e147788a91e-config\") pod \"dnsmasq-dns-675f4bcbfc-945jr\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.355084 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wbd5\" (UniqueName: \"kubernetes.io/projected/63236751-76a1-457c-9885-6e147788a91e-kube-api-access-4wbd5\") pod \"dnsmasq-dns-675f4bcbfc-945jr\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.456957 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wnps\" (UniqueName: \"kubernetes.io/projected/782758d6-63a6-4795-8a8c-9530423e439f-kube-api-access-8wnps\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.457089 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63236751-76a1-457c-9885-6e147788a91e-config\") pod \"dnsmasq-dns-675f4bcbfc-945jr\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.457134 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.457160 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-config\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.457183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wbd5\" (UniqueName: \"kubernetes.io/projected/63236751-76a1-457c-9885-6e147788a91e-kube-api-access-4wbd5\") pod \"dnsmasq-dns-675f4bcbfc-945jr\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.461474 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63236751-76a1-457c-9885-6e147788a91e-config\") pod \"dnsmasq-dns-675f4bcbfc-945jr\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.484473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wbd5\" (UniqueName: \"kubernetes.io/projected/63236751-76a1-457c-9885-6e147788a91e-kube-api-access-4wbd5\") pod \"dnsmasq-dns-675f4bcbfc-945jr\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.502580 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.558758 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.558798 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-config\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.558860 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wnps\" (UniqueName: \"kubernetes.io/projected/782758d6-63a6-4795-8a8c-9530423e439f-kube-api-access-8wnps\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.559595 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.559794 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-config\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.575577 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wnps\" (UniqueName: \"kubernetes.io/projected/782758d6-63a6-4795-8a8c-9530423e439f-kube-api-access-8wnps\") pod \"dnsmasq-dns-78dd6ddcc-d5phf\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:51 crc kubenswrapper[4696]: I1125 10:52:51.607119 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:52:52 crc kubenswrapper[4696]: I1125 10:52:52.112068 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-945jr"] Nov 25 10:52:52 crc kubenswrapper[4696]: I1125 10:52:52.158104 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d5phf"] Nov 25 10:52:52 crc kubenswrapper[4696]: W1125 10:52:52.163414 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod782758d6_63a6_4795_8a8c_9530423e439f.slice/crio-ab0981b1521545b21efd16ad92346285a5a0c2ce95bc3d3bc10987a4f8d4bbcf WatchSource:0}: Error finding container ab0981b1521545b21efd16ad92346285a5a0c2ce95bc3d3bc10987a4f8d4bbcf: Status 404 returned error can't find the container with id ab0981b1521545b21efd16ad92346285a5a0c2ce95bc3d3bc10987a4f8d4bbcf Nov 25 10:52:52 crc kubenswrapper[4696]: I1125 10:52:52.636010 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" event={"ID":"63236751-76a1-457c-9885-6e147788a91e","Type":"ContainerStarted","Data":"ec79792a5397ea384ee9ee61a4074d41190260295d215410cee71c691b2561ed"} Nov 25 10:52:52 crc kubenswrapper[4696]: I1125 10:52:52.637135 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" event={"ID":"782758d6-63a6-4795-8a8c-9530423e439f","Type":"ContainerStarted","Data":"ab0981b1521545b21efd16ad92346285a5a0c2ce95bc3d3bc10987a4f8d4bbcf"} Nov 25 10:52:53 crc kubenswrapper[4696]: I1125 10:52:53.902226 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-945jr"] Nov 25 10:52:53 crc kubenswrapper[4696]: I1125 10:52:53.942773 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rh9xk"] Nov 25 10:52:53 crc kubenswrapper[4696]: I1125 10:52:53.958014 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:53 crc kubenswrapper[4696]: I1125 10:52:53.960116 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rh9xk"] Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.098637 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.098712 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-config\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.098808 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f77s\" (UniqueName: \"kubernetes.io/projected/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-kube-api-access-7f77s\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.200485 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f77s\" (UniqueName: \"kubernetes.io/projected/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-kube-api-access-7f77s\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.200552 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.200579 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-config\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.201568 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-config\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.201878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.266638 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f77s\" (UniqueName: \"kubernetes.io/projected/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-kube-api-access-7f77s\") pod \"dnsmasq-dns-666b6646f7-rh9xk\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.298494 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.350374 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d5phf"] Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.466732 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2mxjc"] Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.468133 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.503407 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2mxjc"] Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.609526 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c987l\" (UniqueName: \"kubernetes.io/projected/b5eaaf49-99f9-446b-a16d-aa792f882507-kube-api-access-c987l\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.609600 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-config\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.609646 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.716079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-config\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.716172 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.716281 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c987l\" (UniqueName: \"kubernetes.io/projected/b5eaaf49-99f9-446b-a16d-aa792f882507-kube-api-access-c987l\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.717595 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-config\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.718181 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.748924 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c987l\" (UniqueName: \"kubernetes.io/projected/b5eaaf49-99f9-446b-a16d-aa792f882507-kube-api-access-c987l\") pod \"dnsmasq-dns-57d769cc4f-2mxjc\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.814870 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:52:54 crc kubenswrapper[4696]: I1125 10:52:54.946727 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rh9xk"] Nov 25 10:52:54 crc kubenswrapper[4696]: W1125 10:52:54.968534 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3425b92_a26b_4077_a2a3_c1ce3dbc4ad8.slice/crio-b54809ee68984f8a2f3731d948cd575aaac7a5be11698328a8d6ea42f4dc7ac5 WatchSource:0}: Error finding container b54809ee68984f8a2f3731d948cd575aaac7a5be11698328a8d6ea42f4dc7ac5: Status 404 returned error can't find the container with id b54809ee68984f8a2f3731d948cd575aaac7a5be11698328a8d6ea42f4dc7ac5 Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.160764 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.168967 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.171989 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-j2rpp" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.172805 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.172945 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.175374 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.175737 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.175396 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.184912 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.198820 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346349 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-server-conf\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346724 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346811 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01da0ce5-e289-4f9d-95d2-922e92053827-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346852 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346867 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346886 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-config-data\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346920 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346946 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh895\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-kube-api-access-bh895\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.346974 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01da0ce5-e289-4f9d-95d2-922e92053827-pod-info\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.347002 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448793 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh895\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-kube-api-access-bh895\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448838 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01da0ce5-e289-4f9d-95d2-922e92053827-pod-info\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448869 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448895 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-server-conf\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448928 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448948 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01da0ce5-e289-4f9d-95d2-922e92053827-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.448987 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.449006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.449029 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-config-data\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.449051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.449072 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.449474 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.456105 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.456531 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-config-data\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.456807 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.457651 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-server-conf\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.458720 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.463461 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.464504 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2mxjc"] Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.467298 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01da0ce5-e289-4f9d-95d2-922e92053827-pod-info\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.469711 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01da0ce5-e289-4f9d-95d2-922e92053827-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.480196 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.515163 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh895\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-kube-api-access-bh895\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.536510 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.630144 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.633609 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.644312 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.644500 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.644679 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.644788 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.644905 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g5b2j" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.645002 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.645092 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.661643 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.722066 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" event={"ID":"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8","Type":"ContainerStarted","Data":"b54809ee68984f8a2f3731d948cd575aaac7a5be11698328a8d6ea42f4dc7ac5"} Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.724336 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" event={"ID":"b5eaaf49-99f9-446b-a16d-aa792f882507","Type":"ContainerStarted","Data":"6d0255020bc24c67baca1744e708f38b079a1c6d34f9b5935f52068b86af0baf"} Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754277 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754357 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754592 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzm8j\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-kube-api-access-lzm8j\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754616 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ff36659-0a3a-4a6f-9f79-414918df363a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754642 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754685 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754716 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754738 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754776 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754908 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.754983 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ff36659-0a3a-4a6f-9f79-414918df363a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.821650 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.855961 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.856001 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.856021 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.856135 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.858899 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.858996 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.859135 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.859159 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.859194 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ff36659-0a3a-4a6f-9f79-414918df363a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.859627 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.859813 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.860053 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.860276 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.860480 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzm8j\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-kube-api-access-lzm8j\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.860601 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ff36659-0a3a-4a6f-9f79-414918df363a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.860718 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.861180 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.863464 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.866399 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ff36659-0a3a-4a6f-9f79-414918df363a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.866753 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ff36659-0a3a-4a6f-9f79-414918df363a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.875962 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzm8j\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-kube-api-access-lzm8j\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.877373 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.890008 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:55 crc kubenswrapper[4696]: I1125 10:52:55.985552 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.578077 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.679536 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.732544 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2ff36659-0a3a-4a6f-9f79-414918df363a","Type":"ContainerStarted","Data":"64019ebf7a097dc66556f3c0d58c6e0e25bdb8e664ccab550551500e85eb5b29"} Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.736838 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01da0ce5-e289-4f9d-95d2-922e92053827","Type":"ContainerStarted","Data":"e646c4f0740ba9eaea56fad2b513d8d83d8b36fab0be6a5bab5d49bfc75650a2"} Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.749148 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.750578 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.756536 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gh2fs" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.756638 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.756908 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.759000 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.768815 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.791175 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880631 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ad9b8866-501c-48e8-a694-5fff74e631fc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880764 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-config-data-default\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880817 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880844 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-kolla-config\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880868 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b8866-501c-48e8-a694-5fff74e631fc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad9b8866-501c-48e8-a694-5fff74e631fc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880946 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.880984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zxjp\" (UniqueName: \"kubernetes.io/projected/ad9b8866-501c-48e8-a694-5fff74e631fc-kube-api-access-6zxjp\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982632 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ad9b8866-501c-48e8-a694-5fff74e631fc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982691 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-config-data-default\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982731 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982755 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b8866-501c-48e8-a694-5fff74e631fc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982769 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-kolla-config\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982785 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad9b8866-501c-48e8-a694-5fff74e631fc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.982862 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zxjp\" (UniqueName: \"kubernetes.io/projected/ad9b8866-501c-48e8-a694-5fff74e631fc-kube-api-access-6zxjp\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.983823 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ad9b8866-501c-48e8-a694-5fff74e631fc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.984471 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-config-data-default\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.984652 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.991465 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-kolla-config\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:56 crc kubenswrapper[4696]: I1125 10:52:56.994092 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad9b8866-501c-48e8-a694-5fff74e631fc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.000334 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b8866-501c-48e8-a694-5fff74e631fc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.002414 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad9b8866-501c-48e8-a694-5fff74e631fc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.024936 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.029328 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zxjp\" (UniqueName: \"kubernetes.io/projected/ad9b8866-501c-48e8-a694-5fff74e631fc-kube-api-access-6zxjp\") pod \"openstack-galera-0\" (UID: \"ad9b8866-501c-48e8-a694-5fff74e631fc\") " pod="openstack/openstack-galera-0" Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.091530 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.567866 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:52:57 crc kubenswrapper[4696]: I1125 10:52:57.759914 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ad9b8866-501c-48e8-a694-5fff74e631fc","Type":"ContainerStarted","Data":"fc3c7199f301e7d28d254f50d1c7dd54dc7bbb856a66a782719dacf1d601bcf9"} Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.018599 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.019935 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.022392 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.022476 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-69tzq" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.022594 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.023247 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.028737 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.108942 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109031 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109077 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c541fdf-4691-4f61-9865-8497d636667d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109101 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h48j\" (UniqueName: \"kubernetes.io/projected/8c541fdf-4691-4f61-9865-8497d636667d-kube-api-access-4h48j\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109127 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109287 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c541fdf-4691-4f61-9865-8497d636667d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109412 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.109470 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c541fdf-4691-4f61-9865-8497d636667d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211125 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211193 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c541fdf-4691-4f61-9865-8497d636667d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211213 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h48j\" (UniqueName: \"kubernetes.io/projected/8c541fdf-4691-4f61-9865-8497d636667d-kube-api-access-4h48j\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211237 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211275 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c541fdf-4691-4f61-9865-8497d636667d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211311 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211333 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c541fdf-4691-4f61-9865-8497d636667d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.211354 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.212195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.212518 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.213727 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c541fdf-4691-4f61-9865-8497d636667d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.214477 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.215546 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c541fdf-4691-4f61-9865-8497d636667d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.227129 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c541fdf-4691-4f61-9865-8497d636667d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.228840 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c541fdf-4691-4f61-9865-8497d636667d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.239454 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h48j\" (UniqueName: \"kubernetes.io/projected/8c541fdf-4691-4f61-9865-8497d636667d-kube-api-access-4h48j\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.251796 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8c541fdf-4691-4f61-9865-8497d636667d\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.338992 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.379883 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.381074 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.384387 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.384604 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.386773 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dcfg9" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.386915 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.515969 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba40268-f540-4ac7-9ae4-14455c81395f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.516014 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqmgl\" (UniqueName: \"kubernetes.io/projected/eba40268-f540-4ac7-9ae4-14455c81395f-kube-api-access-fqmgl\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.516052 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba40268-f540-4ac7-9ae4-14455c81395f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.516093 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/eba40268-f540-4ac7-9ae4-14455c81395f-kolla-config\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.516121 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eba40268-f540-4ac7-9ae4-14455c81395f-config-data\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.617835 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/eba40268-f540-4ac7-9ae4-14455c81395f-kolla-config\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.617891 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eba40268-f540-4ac7-9ae4-14455c81395f-config-data\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.617949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba40268-f540-4ac7-9ae4-14455c81395f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.617967 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqmgl\" (UniqueName: \"kubernetes.io/projected/eba40268-f540-4ac7-9ae4-14455c81395f-kube-api-access-fqmgl\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.618002 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba40268-f540-4ac7-9ae4-14455c81395f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.618801 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/eba40268-f540-4ac7-9ae4-14455c81395f-kolla-config\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.619266 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eba40268-f540-4ac7-9ae4-14455c81395f-config-data\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.621608 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba40268-f540-4ac7-9ae4-14455c81395f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.642213 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqmgl\" (UniqueName: \"kubernetes.io/projected/eba40268-f540-4ac7-9ae4-14455c81395f-kube-api-access-fqmgl\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.642929 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba40268-f540-4ac7-9ae4-14455c81395f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"eba40268-f540-4ac7-9ae4-14455c81395f\") " pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.700241 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:52:58 crc kubenswrapper[4696]: I1125 10:52:58.814008 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:52:59 crc kubenswrapper[4696]: I1125 10:52:59.305513 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:52:59 crc kubenswrapper[4696]: I1125 10:52:59.825448 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c541fdf-4691-4f61-9865-8497d636667d","Type":"ContainerStarted","Data":"3f7e602a29a1f14eb7f81c0e72b9461e89b58a5443032e3db1b45fcae886e189"} Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.413444 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.414438 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.417464 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.420513 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-kp2fm" Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.575636 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577xl\" (UniqueName: \"kubernetes.io/projected/87ef4570-7d70-44f8-9249-62d32b384642-kube-api-access-577xl\") pod \"kube-state-metrics-0\" (UID: \"87ef4570-7d70-44f8-9249-62d32b384642\") " pod="openstack/kube-state-metrics-0" Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.677390 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577xl\" (UniqueName: \"kubernetes.io/projected/87ef4570-7d70-44f8-9249-62d32b384642-kube-api-access-577xl\") pod \"kube-state-metrics-0\" (UID: \"87ef4570-7d70-44f8-9249-62d32b384642\") " pod="openstack/kube-state-metrics-0" Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.701637 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577xl\" (UniqueName: \"kubernetes.io/projected/87ef4570-7d70-44f8-9249-62d32b384642-kube-api-access-577xl\") pod \"kube-state-metrics-0\" (UID: \"87ef4570-7d70-44f8-9249-62d32b384642\") " pod="openstack/kube-state-metrics-0" Nov 25 10:53:00 crc kubenswrapper[4696]: I1125 10:53:00.748948 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:53:03 crc kubenswrapper[4696]: I1125 10:53:03.873294 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"eba40268-f540-4ac7-9ae4-14455c81395f","Type":"ContainerStarted","Data":"4cd907b77bcf70b99c68e7c7f8a8cc1477a80d6c652f5d41298621847417a9e5"} Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.070318 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jz4h7"] Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.117034 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jz4h7"] Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.117141 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.122545 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.122819 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-hgvhr" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.122902 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.134724 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7w2gw"] Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.139809 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.170850 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7w2gw"] Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242532 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-log\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242570 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87dc2400-74ac-4b0b-ae7d-ca62010572d3-scripts\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242601 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-log-ovn\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242622 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dc2400-74ac-4b0b-ae7d-ca62010572d3-combined-ca-bundle\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242652 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xzfn\" (UniqueName: \"kubernetes.io/projected/cb7e6448-16df-4635-84d9-97f6ee770c36-kube-api-access-8xzfn\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242696 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-run\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242715 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-run-ovn\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242744 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-lib\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242773 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb7e6448-16df-4635-84d9-97f6ee770c36-scripts\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242789 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vthjz\" (UniqueName: \"kubernetes.io/projected/87dc2400-74ac-4b0b-ae7d-ca62010572d3-kube-api-access-vthjz\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242806 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-etc-ovs\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242830 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-run\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.242859 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dc2400-74ac-4b0b-ae7d-ca62010572d3-ovn-controller-tls-certs\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.344953 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb7e6448-16df-4635-84d9-97f6ee770c36-scripts\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.345007 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vthjz\" (UniqueName: \"kubernetes.io/projected/87dc2400-74ac-4b0b-ae7d-ca62010572d3-kube-api-access-vthjz\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.345561 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-etc-ovs\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346092 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-run\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346109 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-etc-ovs\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.345599 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-run\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dc2400-74ac-4b0b-ae7d-ca62010572d3-ovn-controller-tls-certs\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346275 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-log\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346303 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87dc2400-74ac-4b0b-ae7d-ca62010572d3-scripts\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346331 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-log-ovn\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346353 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dc2400-74ac-4b0b-ae7d-ca62010572d3-combined-ca-bundle\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346383 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xzfn\" (UniqueName: \"kubernetes.io/projected/cb7e6448-16df-4635-84d9-97f6ee770c36-kube-api-access-8xzfn\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346403 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-run\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346419 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-run-ovn\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346494 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-lib\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346643 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-lib\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.346939 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb7e6448-16df-4635-84d9-97f6ee770c36-scripts\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.347174 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-run\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.347212 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-run-ovn\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.347384 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87dc2400-74ac-4b0b-ae7d-ca62010572d3-var-log-ovn\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.347406 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cb7e6448-16df-4635-84d9-97f6ee770c36-var-log\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.349728 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87dc2400-74ac-4b0b-ae7d-ca62010572d3-scripts\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.366429 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dc2400-74ac-4b0b-ae7d-ca62010572d3-combined-ca-bundle\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.368351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dc2400-74ac-4b0b-ae7d-ca62010572d3-ovn-controller-tls-certs\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.369688 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vthjz\" (UniqueName: \"kubernetes.io/projected/87dc2400-74ac-4b0b-ae7d-ca62010572d3-kube-api-access-vthjz\") pod \"ovn-controller-jz4h7\" (UID: \"87dc2400-74ac-4b0b-ae7d-ca62010572d3\") " pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.380209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xzfn\" (UniqueName: \"kubernetes.io/projected/cb7e6448-16df-4635-84d9-97f6ee770c36-kube-api-access-8xzfn\") pod \"ovn-controller-ovs-7w2gw\" (UID: \"cb7e6448-16df-4635-84d9-97f6ee770c36\") " pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.470938 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.487186 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.936290 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.937947 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.943279 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.943557 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.943592 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.943917 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.944079 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8k7dl" Nov 25 10:53:04 crc kubenswrapper[4696]: I1125 10:53:04.968867 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059152 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbc1467c-cf49-44a9-99eb-c592c38412a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059211 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrjt6\" (UniqueName: \"kubernetes.io/projected/dbc1467c-cf49-44a9-99eb-c592c38412a1-kube-api-access-zrjt6\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059239 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059263 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059296 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc1467c-cf49-44a9-99eb-c592c38412a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059349 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059478 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbc1467c-cf49-44a9-99eb-c592c38412a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.059509 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.160711 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbc1467c-cf49-44a9-99eb-c592c38412a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161175 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrjt6\" (UniqueName: \"kubernetes.io/projected/dbc1467c-cf49-44a9-99eb-c592c38412a1-kube-api-access-zrjt6\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161200 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161231 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161258 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc1467c-cf49-44a9-99eb-c592c38412a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161310 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161352 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbc1467c-cf49-44a9-99eb-c592c38412a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161379 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161548 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbc1467c-cf49-44a9-99eb-c592c38412a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.161914 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.163079 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc1467c-cf49-44a9-99eb-c592c38412a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.163168 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbc1467c-cf49-44a9-99eb-c592c38412a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.166184 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.166251 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.170058 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc1467c-cf49-44a9-99eb-c592c38412a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.179266 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrjt6\" (UniqueName: \"kubernetes.io/projected/dbc1467c-cf49-44a9-99eb-c592c38412a1-kube-api-access-zrjt6\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.192382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbc1467c-cf49-44a9-99eb-c592c38412a1\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:05 crc kubenswrapper[4696]: I1125 10:53:05.268838 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.373578 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.377395 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.380297 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.381315 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-vkgck" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.382284 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.380467 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.388862 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.497874 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.498637 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.498682 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.498704 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.498719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.498748 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blptj\" (UniqueName: \"kubernetes.io/projected/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-kube-api-access-blptj\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.499196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.499234 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-config\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602511 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602534 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602553 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602569 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602598 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blptj\" (UniqueName: \"kubernetes.io/projected/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-kube-api-access-blptj\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602697 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-config\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.602713 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.603157 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.605047 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.605289 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-config\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.606075 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.608923 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.609418 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.623502 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.628745 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blptj\" (UniqueName: \"kubernetes.io/projected/8b4e55fc-7bf4-4a72-b805-b7b2776de33c-kube-api-access-blptj\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.637816 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8b4e55fc-7bf4-4a72-b805-b7b2776de33c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:07 crc kubenswrapper[4696]: I1125 10:53:07.707979 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.394338 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.394879 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7f77s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-rh9xk_openstack(c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.396416 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.425809 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.425996 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wbd5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-945jr_openstack(63236751-76a1-457c-9885-6e147788a91e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.427185 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" podUID="63236751-76a1-457c-9885-6e147788a91e" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.717933 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.718400 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8wnps,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-d5phf_openstack(782758d6-63a6-4795-8a8c-9530423e439f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.719866 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" podUID="782758d6-63a6-4795-8a8c-9530423e439f" Nov 25 10:53:13 crc kubenswrapper[4696]: E1125 10:53:13.950499 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" Nov 25 10:53:14 crc kubenswrapper[4696]: E1125 10:53:14.780503 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 25 10:53:14 crc kubenswrapper[4696]: E1125 10:53:14.780695 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bh895,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(01da0ce5-e289-4f9d-95d2-922e92053827): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:53:14 crc kubenswrapper[4696]: E1125 10:53:14.781914 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" Nov 25 10:53:14 crc kubenswrapper[4696]: E1125 10:53:14.955759 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" Nov 25 10:53:18 crc kubenswrapper[4696]: I1125 10:53:18.991120 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" event={"ID":"782758d6-63a6-4795-8a8c-9530423e439f","Type":"ContainerDied","Data":"ab0981b1521545b21efd16ad92346285a5a0c2ce95bc3d3bc10987a4f8d4bbcf"} Nov 25 10:53:18 crc kubenswrapper[4696]: I1125 10:53:18.991845 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab0981b1521545b21efd16ad92346285a5a0c2ce95bc3d3bc10987a4f8d4bbcf" Nov 25 10:53:18 crc kubenswrapper[4696]: I1125 10:53:18.993311 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" event={"ID":"63236751-76a1-457c-9885-6e147788a91e","Type":"ContainerDied","Data":"ec79792a5397ea384ee9ee61a4074d41190260295d215410cee71c691b2561ed"} Nov 25 10:53:18 crc kubenswrapper[4696]: I1125 10:53:18.993366 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec79792a5397ea384ee9ee61a4074d41190260295d215410cee71c691b2561ed" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.023286 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.028156 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.118377 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wbd5\" (UniqueName: \"kubernetes.io/projected/63236751-76a1-457c-9885-6e147788a91e-kube-api-access-4wbd5\") pod \"63236751-76a1-457c-9885-6e147788a91e\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.118485 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-config\") pod \"782758d6-63a6-4795-8a8c-9530423e439f\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.118552 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-dns-svc\") pod \"782758d6-63a6-4795-8a8c-9530423e439f\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.118594 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63236751-76a1-457c-9885-6e147788a91e-config\") pod \"63236751-76a1-457c-9885-6e147788a91e\" (UID: \"63236751-76a1-457c-9885-6e147788a91e\") " Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.118790 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wnps\" (UniqueName: \"kubernetes.io/projected/782758d6-63a6-4795-8a8c-9530423e439f-kube-api-access-8wnps\") pod \"782758d6-63a6-4795-8a8c-9530423e439f\" (UID: \"782758d6-63a6-4795-8a8c-9530423e439f\") " Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.119133 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "782758d6-63a6-4795-8a8c-9530423e439f" (UID: "782758d6-63a6-4795-8a8c-9530423e439f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.119166 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-config" (OuterVolumeSpecName: "config") pod "782758d6-63a6-4795-8a8c-9530423e439f" (UID: "782758d6-63a6-4795-8a8c-9530423e439f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.119133 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63236751-76a1-457c-9885-6e147788a91e-config" (OuterVolumeSpecName: "config") pod "63236751-76a1-457c-9885-6e147788a91e" (UID: "63236751-76a1-457c-9885-6e147788a91e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.120590 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.120627 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782758d6-63a6-4795-8a8c-9530423e439f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.120646 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63236751-76a1-457c-9885-6e147788a91e-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.125530 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63236751-76a1-457c-9885-6e147788a91e-kube-api-access-4wbd5" (OuterVolumeSpecName: "kube-api-access-4wbd5") pod "63236751-76a1-457c-9885-6e147788a91e" (UID: "63236751-76a1-457c-9885-6e147788a91e"). InnerVolumeSpecName "kube-api-access-4wbd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.125949 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782758d6-63a6-4795-8a8c-9530423e439f-kube-api-access-8wnps" (OuterVolumeSpecName: "kube-api-access-8wnps") pod "782758d6-63a6-4795-8a8c-9530423e439f" (UID: "782758d6-63a6-4795-8a8c-9530423e439f"). InnerVolumeSpecName "kube-api-access-8wnps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.222981 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wbd5\" (UniqueName: \"kubernetes.io/projected/63236751-76a1-457c-9885-6e147788a91e-kube-api-access-4wbd5\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:19 crc kubenswrapper[4696]: I1125 10:53:19.223302 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wnps\" (UniqueName: \"kubernetes.io/projected/782758d6-63a6-4795-8a8c-9530423e439f-kube-api-access-8wnps\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4696]: I1125 10:53:20.000345 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-945jr" Nov 25 10:53:20 crc kubenswrapper[4696]: I1125 10:53:20.000367 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d5phf" Nov 25 10:53:20 crc kubenswrapper[4696]: I1125 10:53:20.080056 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-945jr"] Nov 25 10:53:20 crc kubenswrapper[4696]: I1125 10:53:20.112774 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-945jr"] Nov 25 10:53:20 crc kubenswrapper[4696]: I1125 10:53:20.127114 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d5phf"] Nov 25 10:53:20 crc kubenswrapper[4696]: I1125 10:53:20.130647 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d5phf"] Nov 25 10:53:22 crc kubenswrapper[4696]: I1125 10:53:22.058055 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63236751-76a1-457c-9885-6e147788a91e" path="/var/lib/kubelet/pods/63236751-76a1-457c-9885-6e147788a91e/volumes" Nov 25 10:53:22 crc kubenswrapper[4696]: I1125 10:53:22.058773 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782758d6-63a6-4795-8a8c-9530423e439f" path="/var/lib/kubelet/pods/782758d6-63a6-4795-8a8c-9530423e439f/volumes" Nov 25 10:53:26 crc kubenswrapper[4696]: I1125 10:53:26.996362 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:53:27 crc kubenswrapper[4696]: I1125 10:53:27.057724 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jz4h7"] Nov 25 10:53:27 crc kubenswrapper[4696]: I1125 10:53:27.075151 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"87ef4570-7d70-44f8-9249-62d32b384642","Type":"ContainerStarted","Data":"7fda7bf0de2bcc7e7449a0d17f427ff3034d4e190d4ece264a28cf631af9d6da"} Nov 25 10:53:27 crc kubenswrapper[4696]: E1125 10:53:27.171227 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Nov 25 10:53:27 crc kubenswrapper[4696]: E1125 10:53:27.171516 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n65ch5hbfh59chbdh97h5d8h69h5c5h77h9dh5d8h5f8h694h5bch5c5h5b9h68hf4h646hbdh6dh579h5bdh59bh68dhc7hch5f7h587h65bhf5q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(eba40268-f540-4ac7-9ae4-14455c81395f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:53:27 crc kubenswrapper[4696]: E1125 10:53:27.172840 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="eba40268-f540-4ac7-9ae4-14455c81395f" Nov 25 10:53:27 crc kubenswrapper[4696]: E1125 10:53:27.759644 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Nov 25 10:53:27 crc kubenswrapper[4696]: E1125 10:53:27.760822 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6zxjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(ad9b8866-501c-48e8-a694-5fff74e631fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:53:27 crc kubenswrapper[4696]: E1125 10:53:27.762252 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="ad9b8866-501c-48e8-a694-5fff74e631fc" Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.055924 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:53:28 crc kubenswrapper[4696]: W1125 10:53:28.063734 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbc1467c_cf49_44a9_99eb_c592c38412a1.slice/crio-9348cba1681d377dda8170a130fff27dcdd62b21af815b4ddd138b7e55049ea5 WatchSource:0}: Error finding container 9348cba1681d377dda8170a130fff27dcdd62b21af815b4ddd138b7e55049ea5: Status 404 returned error can't find the container with id 9348cba1681d377dda8170a130fff27dcdd62b21af815b4ddd138b7e55049ea5 Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.100738 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbc1467c-cf49-44a9-99eb-c592c38412a1","Type":"ContainerStarted","Data":"9348cba1681d377dda8170a130fff27dcdd62b21af815b4ddd138b7e55049ea5"} Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.119056 4696 generic.go:334] "Generic (PLEG): container finished" podID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerID="a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5" exitCode=0 Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.119159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" event={"ID":"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8","Type":"ContainerDied","Data":"a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5"} Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.124136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01da0ce5-e289-4f9d-95d2-922e92053827","Type":"ContainerStarted","Data":"c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b"} Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.127292 4696 generic.go:334] "Generic (PLEG): container finished" podID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerID="a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c" exitCode=0 Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.127370 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" event={"ID":"b5eaaf49-99f9-446b-a16d-aa792f882507","Type":"ContainerDied","Data":"a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c"} Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.135890 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c541fdf-4691-4f61-9865-8497d636667d","Type":"ContainerStarted","Data":"8f320077157e9b7f8b5a389f5b563d35142ede3d9938f4c0891ad83c33d4de0d"} Nov 25 10:53:28 crc kubenswrapper[4696]: I1125 10:53:28.141849 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jz4h7" event={"ID":"87dc2400-74ac-4b0b-ae7d-ca62010572d3","Type":"ContainerStarted","Data":"abbd7ba51ae231b2870770c4fbf0a3b6c02333897aae51bd6b5cf0d2b631e8d0"} Nov 25 10:53:28 crc kubenswrapper[4696]: E1125 10:53:28.143196 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="eba40268-f540-4ac7-9ae4-14455c81395f" Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.007067 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7w2gw"] Nov 25 10:53:29 crc kubenswrapper[4696]: W1125 10:53:29.073955 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb7e6448_16df_4635_84d9_97f6ee770c36.slice/crio-ac193700b388dcbc3c55de2c37cefadb9a4790be81e92a73036bcf70c3f1f0c8 WatchSource:0}: Error finding container ac193700b388dcbc3c55de2c37cefadb9a4790be81e92a73036bcf70c3f1f0c8: Status 404 returned error can't find the container with id ac193700b388dcbc3c55de2c37cefadb9a4790be81e92a73036bcf70c3f1f0c8 Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.124596 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.154157 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" event={"ID":"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8","Type":"ContainerStarted","Data":"aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994"} Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.154541 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.157041 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7w2gw" event={"ID":"cb7e6448-16df-4635-84d9-97f6ee770c36","Type":"ContainerStarted","Data":"ac193700b388dcbc3c55de2c37cefadb9a4790be81e92a73036bcf70c3f1f0c8"} Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.159626 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" event={"ID":"b5eaaf49-99f9-446b-a16d-aa792f882507","Type":"ContainerStarted","Data":"88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532"} Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.159713 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.161296 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2ff36659-0a3a-4a6f-9f79-414918df363a","Type":"ContainerStarted","Data":"fa9cbf49307a6d897269a72edeb254cafcbb7dab2a48856998a38a4b27815fba"} Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.165545 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ad9b8866-501c-48e8-a694-5fff74e631fc","Type":"ContainerStarted","Data":"c6bcd342422b7ef06f17baed5e485a64d54cbcd0e6a78c32031e2604970a3583"} Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.174759 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" podStartSLOduration=-9223372000.680035 podStartE2EDuration="36.174741274s" podCreationTimestamp="2025-11-25 10:52:53 +0000 UTC" firstStartedPulling="2025-11-25 10:52:54.97557879 +0000 UTC m=+971.789195377" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:29.173940958 +0000 UTC m=+1005.987557565" watchObservedRunningTime="2025-11-25 10:53:29.174741274 +0000 UTC m=+1005.988357861" Nov 25 10:53:29 crc kubenswrapper[4696]: I1125 10:53:29.241637 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" podStartSLOduration=3.607825899 podStartE2EDuration="35.241616443s" podCreationTimestamp="2025-11-25 10:52:54 +0000 UTC" firstStartedPulling="2025-11-25 10:52:55.512255248 +0000 UTC m=+972.325871835" lastFinishedPulling="2025-11-25 10:53:27.146045752 +0000 UTC m=+1003.959662379" observedRunningTime="2025-11-25 10:53:29.234583813 +0000 UTC m=+1006.048200400" watchObservedRunningTime="2025-11-25 10:53:29.241616443 +0000 UTC m=+1006.055233030" Nov 25 10:53:30 crc kubenswrapper[4696]: I1125 10:53:30.172976 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8b4e55fc-7bf4-4a72-b805-b7b2776de33c","Type":"ContainerStarted","Data":"7f7d5356c1cf945931adb53bed72f997e28149f6506078807b016cc15667f8ab"} Nov 25 10:53:30 crc kubenswrapper[4696]: I1125 10:53:30.802144 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:53:30 crc kubenswrapper[4696]: I1125 10:53:30.802493 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.192177 4696 generic.go:334] "Generic (PLEG): container finished" podID="ad9b8866-501c-48e8-a694-5fff74e631fc" containerID="c6bcd342422b7ef06f17baed5e485a64d54cbcd0e6a78c32031e2604970a3583" exitCode=0 Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.192253 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ad9b8866-501c-48e8-a694-5fff74e631fc","Type":"ContainerDied","Data":"c6bcd342422b7ef06f17baed5e485a64d54cbcd0e6a78c32031e2604970a3583"} Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.221346 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbc1467c-cf49-44a9-99eb-c592c38412a1","Type":"ContainerStarted","Data":"a8b84fe75521a80f1bdcc756d2c4abb612b4ede229268759e2c8e9c978df652d"} Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.228440 4696 generic.go:334] "Generic (PLEG): container finished" podID="8c541fdf-4691-4f61-9865-8497d636667d" containerID="8f320077157e9b7f8b5a389f5b563d35142ede3d9938f4c0891ad83c33d4de0d" exitCode=0 Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.228526 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c541fdf-4691-4f61-9865-8497d636667d","Type":"ContainerDied","Data":"8f320077157e9b7f8b5a389f5b563d35142ede3d9938f4c0891ad83c33d4de0d"} Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.233575 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"87ef4570-7d70-44f8-9249-62d32b384642","Type":"ContainerStarted","Data":"8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9"} Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.234493 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:53:32 crc kubenswrapper[4696]: I1125 10:53:32.258229 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=27.336281157 podStartE2EDuration="32.258211758s" podCreationTimestamp="2025-11-25 10:53:00 +0000 UTC" firstStartedPulling="2025-11-25 10:53:26.990415318 +0000 UTC m=+1003.804031905" lastFinishedPulling="2025-11-25 10:53:31.912345919 +0000 UTC m=+1008.725962506" observedRunningTime="2025-11-25 10:53:32.248821615 +0000 UTC m=+1009.062438202" watchObservedRunningTime="2025-11-25 10:53:32.258211758 +0000 UTC m=+1009.071828345" Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.243278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ad9b8866-501c-48e8-a694-5fff74e631fc","Type":"ContainerStarted","Data":"bf653f6568dca8b8523ed8a52b4fe44e779b6f88b0dd4ff9c55014c48d173289"} Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.248613 4696 generic.go:334] "Generic (PLEG): container finished" podID="cb7e6448-16df-4635-84d9-97f6ee770c36" containerID="4d26a75e5594fae0955caaa0127aaa69d3515fd268ed835e90813c883feea1d9" exitCode=0 Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.248689 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7w2gw" event={"ID":"cb7e6448-16df-4635-84d9-97f6ee770c36","Type":"ContainerDied","Data":"4d26a75e5594fae0955caaa0127aaa69d3515fd268ed835e90813c883feea1d9"} Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.250712 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8b4e55fc-7bf4-4a72-b805-b7b2776de33c","Type":"ContainerStarted","Data":"b7cdca2fe189c47d8313aaec0ffa01b5db4550bacd85dd1f3b1061b55cdbc2f7"} Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.268252 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371998.586538 podStartE2EDuration="38.268237823s" podCreationTimestamp="2025-11-25 10:52:55 +0000 UTC" firstStartedPulling="2025-11-25 10:52:57.582896212 +0000 UTC m=+974.396512799" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:33.26527354 +0000 UTC m=+1010.078890127" watchObservedRunningTime="2025-11-25 10:53:33.268237823 +0000 UTC m=+1010.081854410" Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.271592 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c541fdf-4691-4f61-9865-8497d636667d","Type":"ContainerStarted","Data":"5aaa1ee597a7f124d48e0c9f7d0d60539b2946dfa27cd99c86c73aedcc4cd68f"} Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.274375 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jz4h7" event={"ID":"87dc2400-74ac-4b0b-ae7d-ca62010572d3","Type":"ContainerStarted","Data":"c879a36dafe556d208b1989cf521282e16b406366b104e0bf934918aa0455fc4"} Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.274551 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jz4h7" Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.308183 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jz4h7" podStartSLOduration=24.483385546 podStartE2EDuration="29.308166761s" podCreationTimestamp="2025-11-25 10:53:04 +0000 UTC" firstStartedPulling="2025-11-25 10:53:27.070457029 +0000 UTC m=+1003.884073616" lastFinishedPulling="2025-11-25 10:53:31.895238244 +0000 UTC m=+1008.708854831" observedRunningTime="2025-11-25 10:53:33.298148718 +0000 UTC m=+1010.111765305" watchObservedRunningTime="2025-11-25 10:53:33.308166761 +0000 UTC m=+1010.121783348" Nov 25 10:53:33 crc kubenswrapper[4696]: I1125 10:53:33.323024 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.444531972 podStartE2EDuration="37.323005345s" podCreationTimestamp="2025-11-25 10:52:56 +0000 UTC" firstStartedPulling="2025-11-25 10:52:58.826711474 +0000 UTC m=+975.640328061" lastFinishedPulling="2025-11-25 10:53:27.705184837 +0000 UTC m=+1004.518801434" observedRunningTime="2025-11-25 10:53:33.315329526 +0000 UTC m=+1010.128946113" watchObservedRunningTime="2025-11-25 10:53:33.323005345 +0000 UTC m=+1010.136621932" Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.291509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7w2gw" event={"ID":"cb7e6448-16df-4635-84d9-97f6ee770c36","Type":"ContainerStarted","Data":"6a82772270a083700e92017bdb4c6c44472a76872a016a27063104fb8be21031"} Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.291863 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7w2gw" event={"ID":"cb7e6448-16df-4635-84d9-97f6ee770c36","Type":"ContainerStarted","Data":"d8f5b8db42ff2dfda7ec15c0589417de1536f256322697ff3f299f5f2a753b74"} Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.292733 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.292789 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.301497 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.319914 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7w2gw" podStartSLOduration=27.483242778 podStartE2EDuration="30.31989544s" podCreationTimestamp="2025-11-25 10:53:04 +0000 UTC" firstStartedPulling="2025-11-25 10:53:29.076248545 +0000 UTC m=+1005.889865122" lastFinishedPulling="2025-11-25 10:53:31.912901197 +0000 UTC m=+1008.726517784" observedRunningTime="2025-11-25 10:53:34.308717131 +0000 UTC m=+1011.122333788" watchObservedRunningTime="2025-11-25 10:53:34.31989544 +0000 UTC m=+1011.133512027" Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.816878 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:53:34 crc kubenswrapper[4696]: I1125 10:53:34.886106 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rh9xk"] Nov 25 10:53:35 crc kubenswrapper[4696]: I1125 10:53:35.297655 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerName="dnsmasq-dns" containerID="cri-o://aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994" gracePeriod=10 Nov 25 10:53:35 crc kubenswrapper[4696]: E1125 10:53:35.311367 4696 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.38:41012->38.102.83.38:38483: write tcp 38.102.83.38:41012->38.102.83.38:38483: write: connection reset by peer Nov 25 10:53:35 crc kubenswrapper[4696]: E1125 10:53:35.396943 4696 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.38:41014->38.102.83.38:38483: write tcp 38.102.83.38:41014->38.102.83.38:38483: write: connection reset by peer Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.154742 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.253142 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-dns-svc\") pod \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.253184 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-config\") pod \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.253222 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f77s\" (UniqueName: \"kubernetes.io/projected/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-kube-api-access-7f77s\") pod \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\" (UID: \"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8\") " Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.259153 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-kube-api-access-7f77s" (OuterVolumeSpecName: "kube-api-access-7f77s") pod "c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" (UID: "c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8"). InnerVolumeSpecName "kube-api-access-7f77s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.288721 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-config" (OuterVolumeSpecName: "config") pod "c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" (UID: "c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.294390 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" (UID: "c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.307939 4696 generic.go:334] "Generic (PLEG): container finished" podID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerID="aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994" exitCode=0 Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.308015 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.308025 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" event={"ID":"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8","Type":"ContainerDied","Data":"aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994"} Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.308069 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rh9xk" event={"ID":"c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8","Type":"ContainerDied","Data":"b54809ee68984f8a2f3731d948cd575aaac7a5be11698328a8d6ea42f4dc7ac5"} Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.308093 4696 scope.go:117] "RemoveContainer" containerID="aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.312049 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8b4e55fc-7bf4-4a72-b805-b7b2776de33c","Type":"ContainerStarted","Data":"6e1f24e31bbadd213e931d631b62d76b1d12ba577fa567eb66d707ffd4eddba8"} Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.328471 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbc1467c-cf49-44a9-99eb-c592c38412a1","Type":"ContainerStarted","Data":"9c4b50104afa4c348246fb1d418cc4070ad4ce5816c3112f875034de577a6d3e"} Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.333503 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=23.720481689 podStartE2EDuration="30.333489999s" podCreationTimestamp="2025-11-25 10:53:06 +0000 UTC" firstStartedPulling="2025-11-25 10:53:29.279048213 +0000 UTC m=+1006.092664800" lastFinishedPulling="2025-11-25 10:53:35.892056523 +0000 UTC m=+1012.705673110" observedRunningTime="2025-11-25 10:53:36.328612556 +0000 UTC m=+1013.142229143" watchObservedRunningTime="2025-11-25 10:53:36.333489999 +0000 UTC m=+1013.147106586" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.337836 4696 scope.go:117] "RemoveContainer" containerID="a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.346250 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.467713907 podStartE2EDuration="33.346235167s" podCreationTimestamp="2025-11-25 10:53:03 +0000 UTC" firstStartedPulling="2025-11-25 10:53:28.068766609 +0000 UTC m=+1004.882383196" lastFinishedPulling="2025-11-25 10:53:35.947287869 +0000 UTC m=+1012.760904456" observedRunningTime="2025-11-25 10:53:36.343113409 +0000 UTC m=+1013.156729996" watchObservedRunningTime="2025-11-25 10:53:36.346235167 +0000 UTC m=+1013.159851754" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.354745 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.354774 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.354785 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f77s\" (UniqueName: \"kubernetes.io/projected/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8-kube-api-access-7f77s\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.368877 4696 scope.go:117] "RemoveContainer" containerID="aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994" Nov 25 10:53:36 crc kubenswrapper[4696]: E1125 10:53:36.369290 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994\": container with ID starting with aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994 not found: ID does not exist" containerID="aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.369319 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994"} err="failed to get container status \"aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994\": rpc error: code = NotFound desc = could not find container \"aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994\": container with ID starting with aeff75fb2c3f9f0fe28a0a68efa8b2a3b841ff0f411999f6264c55b8d0cc7994 not found: ID does not exist" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.369339 4696 scope.go:117] "RemoveContainer" containerID="a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5" Nov 25 10:53:36 crc kubenswrapper[4696]: E1125 10:53:36.369615 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5\": container with ID starting with a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5 not found: ID does not exist" containerID="a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.369648 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5"} err="failed to get container status \"a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5\": rpc error: code = NotFound desc = could not find container \"a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5\": container with ID starting with a2b56aa3cdd0c71a4aafa5b81ba05e858c2d0165044c747e7714854c7eba06b5 not found: ID does not exist" Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.373360 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rh9xk"] Nov 25 10:53:36 crc kubenswrapper[4696]: I1125 10:53:36.378552 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rh9xk"] Nov 25 10:53:37 crc kubenswrapper[4696]: I1125 10:53:37.091710 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 10:53:37 crc kubenswrapper[4696]: I1125 10:53:37.092024 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 10:53:37 crc kubenswrapper[4696]: I1125 10:53:37.709060 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:37 crc kubenswrapper[4696]: I1125 10:53:37.709095 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:37 crc kubenswrapper[4696]: I1125 10:53:37.746591 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.060768 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" path="/var/lib/kubelet/pods/c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8/volumes" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.269944 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.340042 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.340099 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.357375 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.419476 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.463732 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.589085 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.661530 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-qld29"] Nov 25 10:53:38 crc kubenswrapper[4696]: E1125 10:53:38.661864 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerName="init" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.661880 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerName="init" Nov 25 10:53:38 crc kubenswrapper[4696]: E1125 10:53:38.661910 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerName="dnsmasq-dns" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.661918 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerName="dnsmasq-dns" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.662047 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3425b92-a26b-4077-a2a3-c1ce3dbc4ad8" containerName="dnsmasq-dns" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.662533 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.664933 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.714421 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzv7r\" (UniqueName: \"kubernetes.io/projected/d900f0ea-8197-4383-bdbb-666b60db9f19-kube-api-access-wzv7r\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.715328 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d900f0ea-8197-4383-bdbb-666b60db9f19-ovs-rundir\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.715402 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d900f0ea-8197-4383-bdbb-666b60db9f19-ovn-rundir\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.715555 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d900f0ea-8197-4383-bdbb-666b60db9f19-config\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.715696 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d900f0ea-8197-4383-bdbb-666b60db9f19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.715743 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d900f0ea-8197-4383-bdbb-666b60db9f19-combined-ca-bundle\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.733279 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-87wk4"] Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.735812 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.738428 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.762552 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-87wk4"] Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.770850 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-qld29"] Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817239 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86s5b\" (UniqueName: \"kubernetes.io/projected/eb781fb7-f7dc-468e-b326-38f1e5909127-kube-api-access-86s5b\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817289 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817308 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-config\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817330 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d900f0ea-8197-4383-bdbb-666b60db9f19-config\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817372 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d900f0ea-8197-4383-bdbb-666b60db9f19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817387 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d900f0ea-8197-4383-bdbb-666b60db9f19-combined-ca-bundle\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzv7r\" (UniqueName: \"kubernetes.io/projected/d900f0ea-8197-4383-bdbb-666b60db9f19-kube-api-access-wzv7r\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817447 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d900f0ea-8197-4383-bdbb-666b60db9f19-ovs-rundir\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d900f0ea-8197-4383-bdbb-666b60db9f19-ovn-rundir\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.817488 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.818230 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d900f0ea-8197-4383-bdbb-666b60db9f19-config\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.821967 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d900f0ea-8197-4383-bdbb-666b60db9f19-ovs-rundir\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.822056 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d900f0ea-8197-4383-bdbb-666b60db9f19-ovn-rundir\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.824508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d900f0ea-8197-4383-bdbb-666b60db9f19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.839849 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d900f0ea-8197-4383-bdbb-666b60db9f19-combined-ca-bundle\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.855523 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzv7r\" (UniqueName: \"kubernetes.io/projected/d900f0ea-8197-4383-bdbb-666b60db9f19-kube-api-access-wzv7r\") pod \"ovn-controller-metrics-qld29\" (UID: \"d900f0ea-8197-4383-bdbb-666b60db9f19\") " pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.919304 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.919546 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86s5b\" (UniqueName: \"kubernetes.io/projected/eb781fb7-f7dc-468e-b326-38f1e5909127-kube-api-access-86s5b\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.919579 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.919596 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-config\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.920252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.920424 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-config\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.920986 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.935960 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86s5b\" (UniqueName: \"kubernetes.io/projected/eb781fb7-f7dc-468e-b326-38f1e5909127-kube-api-access-86s5b\") pod \"dnsmasq-dns-5bf47b49b7-87wk4\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:38 crc kubenswrapper[4696]: I1125 10:53:38.988212 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-qld29" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.064865 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.144444 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-87wk4"] Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.173355 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-tppmp"] Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.174687 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.177873 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.217841 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tppmp"] Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.225302 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg88r\" (UniqueName: \"kubernetes.io/projected/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-kube-api-access-fg88r\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.225373 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.225399 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-config\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.225427 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.225457 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-dns-svc\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.326954 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg88r\" (UniqueName: \"kubernetes.io/projected/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-kube-api-access-fg88r\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.327027 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.327051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-config\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.327076 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.327105 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-dns-svc\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.327964 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-dns-svc\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.328876 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.329506 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.329570 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-config\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.331735 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.359596 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg88r\" (UniqueName: \"kubernetes.io/projected/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-kube-api-access-fg88r\") pod \"dnsmasq-dns-8554648995-tppmp\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.374985 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.434925 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 10:53:39 crc kubenswrapper[4696]: W1125 10:53:39.494882 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd900f0ea_8197_4383_bdbb_666b60db9f19.slice/crio-0ef3d76b20349f53114b2d6155c4fb37ef4a6bc6c151169ec233dd58df4c5711 WatchSource:0}: Error finding container 0ef3d76b20349f53114b2d6155c4fb37ef4a6bc6c151169ec233dd58df4c5711: Status 404 returned error can't find the container with id 0ef3d76b20349f53114b2d6155c4fb37ef4a6bc6c151169ec233dd58df4c5711 Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.495388 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.495653 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-qld29"] Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.545059 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.694348 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-87wk4"] Nov 25 10:53:39 crc kubenswrapper[4696]: W1125 10:53:39.729776 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb781fb7_f7dc_468e_b326_38f1e5909127.slice/crio-3bef3071d1016e0b39d485fafea977064fa0cd807bc2574b9017087388112c78 WatchSource:0}: Error finding container 3bef3071d1016e0b39d485fafea977064fa0cd807bc2574b9017087388112c78: Status 404 returned error can't find the container with id 3bef3071d1016e0b39d485fafea977064fa0cd807bc2574b9017087388112c78 Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.845877 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.848246 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.852102 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.852386 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.852545 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-2nwkv" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.852725 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 10:53:39 crc kubenswrapper[4696]: I1125 10:53:39.875212 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.015146 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tppmp"] Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.043838 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.043905 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e0d794d3-7682-4248-affe-b53f4f34f264-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.043947 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.043970 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0d794d3-7682-4248-affe-b53f4f34f264-scripts\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.044013 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0d794d3-7682-4248-affe-b53f4f34f264-config\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.044743 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkg8h\" (UniqueName: \"kubernetes.io/projected/e0d794d3-7682-4248-affe-b53f4f34f264-kube-api-access-qkg8h\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.044910 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: W1125 10:53:40.065747 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda67dc117_d4d1_4cf3_bd47_0683bfc7b108.slice/crio-0c108cf5f9793277ac0b45699a263c53b4beea4515b27b4a590fcf410f3fa8a6 WatchSource:0}: Error finding container 0c108cf5f9793277ac0b45699a263c53b4beea4515b27b4a590fcf410f3fa8a6: Status 404 returned error can't find the container with id 0c108cf5f9793277ac0b45699a263c53b4beea4515b27b4a590fcf410f3fa8a6 Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147216 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147696 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e0d794d3-7682-4248-affe-b53f4f34f264-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147764 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0d794d3-7682-4248-affe-b53f4f34f264-scripts\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147803 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0d794d3-7682-4248-affe-b53f4f34f264-config\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147844 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkg8h\" (UniqueName: \"kubernetes.io/projected/e0d794d3-7682-4248-affe-b53f4f34f264-kube-api-access-qkg8h\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.147892 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.149392 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e0d794d3-7682-4248-affe-b53f4f34f264-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.150619 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0d794d3-7682-4248-affe-b53f4f34f264-config\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.152176 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.153561 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.156696 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0d794d3-7682-4248-affe-b53f4f34f264-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.156776 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0d794d3-7682-4248-affe-b53f4f34f264-scripts\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.165490 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkg8h\" (UniqueName: \"kubernetes.io/projected/e0d794d3-7682-4248-affe-b53f4f34f264-kube-api-access-qkg8h\") pod \"ovn-northd-0\" (UID: \"e0d794d3-7682-4248-affe-b53f4f34f264\") " pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.225847 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.386505 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-qld29" event={"ID":"d900f0ea-8197-4383-bdbb-666b60db9f19","Type":"ContainerStarted","Data":"dbcb11b3cf5e3d756dfecd5e579423bb7cd95b15e229cfdc8355d9cd3527424d"} Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.386547 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-qld29" event={"ID":"d900f0ea-8197-4383-bdbb-666b60db9f19","Type":"ContainerStarted","Data":"0ef3d76b20349f53114b2d6155c4fb37ef4a6bc6c151169ec233dd58df4c5711"} Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.389466 4696 generic.go:334] "Generic (PLEG): container finished" podID="eb781fb7-f7dc-468e-b326-38f1e5909127" containerID="079dbf94d4469182918b956d42ba6d8017880b51a05f8fcfef522cea107e0a68" exitCode=0 Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.389522 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" event={"ID":"eb781fb7-f7dc-468e-b326-38f1e5909127","Type":"ContainerDied","Data":"079dbf94d4469182918b956d42ba6d8017880b51a05f8fcfef522cea107e0a68"} Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.389566 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" event={"ID":"eb781fb7-f7dc-468e-b326-38f1e5909127","Type":"ContainerStarted","Data":"3bef3071d1016e0b39d485fafea977064fa0cd807bc2574b9017087388112c78"} Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.391383 4696 generic.go:334] "Generic (PLEG): container finished" podID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerID="9eff8cb02b324b2fd51b27e80202509f61a1cc0a23b3ea7340663eba2deff5d8" exitCode=0 Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.401434 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tppmp" event={"ID":"a67dc117-d4d1-4cf3-bd47-0683bfc7b108","Type":"ContainerDied","Data":"9eff8cb02b324b2fd51b27e80202509f61a1cc0a23b3ea7340663eba2deff5d8"} Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.401489 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tppmp" event={"ID":"a67dc117-d4d1-4cf3-bd47-0683bfc7b108","Type":"ContainerStarted","Data":"0c108cf5f9793277ac0b45699a263c53b4beea4515b27b4a590fcf410f3fa8a6"} Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.437425 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-qld29" podStartSLOduration=2.437404155 podStartE2EDuration="2.437404155s" podCreationTimestamp="2025-11-25 10:53:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:40.418460252 +0000 UTC m=+1017.232076839" watchObservedRunningTime="2025-11-25 10:53:40.437404155 +0000 UTC m=+1017.251020742" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.567387 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:53:40 crc kubenswrapper[4696]: W1125 10:53:40.577092 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0d794d3_7682_4248_affe_b53f4f34f264.slice/crio-22d13e22c3f0add01393c5ff377e79d0d1bba752a3b7b7763341bc329a832da4 WatchSource:0}: Error finding container 22d13e22c3f0add01393c5ff377e79d0d1bba752a3b7b7763341bc329a832da4: Status 404 returned error can't find the container with id 22d13e22c3f0add01393c5ff377e79d0d1bba752a3b7b7763341bc329a832da4 Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.712781 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.754250 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.859437 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-ovsdbserver-nb\") pod \"eb781fb7-f7dc-468e-b326-38f1e5909127\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.859572 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-dns-svc\") pod \"eb781fb7-f7dc-468e-b326-38f1e5909127\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.859700 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86s5b\" (UniqueName: \"kubernetes.io/projected/eb781fb7-f7dc-468e-b326-38f1e5909127-kube-api-access-86s5b\") pod \"eb781fb7-f7dc-468e-b326-38f1e5909127\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.859785 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-config\") pod \"eb781fb7-f7dc-468e-b326-38f1e5909127\" (UID: \"eb781fb7-f7dc-468e-b326-38f1e5909127\") " Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.866624 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb781fb7-f7dc-468e-b326-38f1e5909127-kube-api-access-86s5b" (OuterVolumeSpecName: "kube-api-access-86s5b") pod "eb781fb7-f7dc-468e-b326-38f1e5909127" (UID: "eb781fb7-f7dc-468e-b326-38f1e5909127"). InnerVolumeSpecName "kube-api-access-86s5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.881353 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb781fb7-f7dc-468e-b326-38f1e5909127" (UID: "eb781fb7-f7dc-468e-b326-38f1e5909127"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.891043 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-config" (OuterVolumeSpecName: "config") pod "eb781fb7-f7dc-468e-b326-38f1e5909127" (UID: "eb781fb7-f7dc-468e-b326-38f1e5909127"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.895919 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb781fb7-f7dc-468e-b326-38f1e5909127" (UID: "eb781fb7-f7dc-468e-b326-38f1e5909127"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.962610 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.962902 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.962983 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86s5b\" (UniqueName: \"kubernetes.io/projected/eb781fb7-f7dc-468e-b326-38f1e5909127-kube-api-access-86s5b\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:40 crc kubenswrapper[4696]: I1125 10:53:40.963076 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb781fb7-f7dc-468e-b326-38f1e5909127-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.400409 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" event={"ID":"eb781fb7-f7dc-468e-b326-38f1e5909127","Type":"ContainerDied","Data":"3bef3071d1016e0b39d485fafea977064fa0cd807bc2574b9017087388112c78"} Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.400478 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-87wk4" Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.400484 4696 scope.go:117] "RemoveContainer" containerID="079dbf94d4469182918b956d42ba6d8017880b51a05f8fcfef522cea107e0a68" Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.406952 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tppmp" event={"ID":"a67dc117-d4d1-4cf3-bd47-0683bfc7b108","Type":"ContainerStarted","Data":"08d01a7c885b90c8624821d4a7447527f42bfcd77f3fddba9bbb010fa6bce2e3"} Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.407049 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.430596 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e0d794d3-7682-4248-affe-b53f4f34f264","Type":"ContainerStarted","Data":"22d13e22c3f0add01393c5ff377e79d0d1bba752a3b7b7763341bc329a832da4"} Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.430640 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-tppmp" podStartSLOduration=2.430613745 podStartE2EDuration="2.430613745s" podCreationTimestamp="2025-11-25 10:53:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:41.428092826 +0000 UTC m=+1018.241709413" watchObservedRunningTime="2025-11-25 10:53:41.430613745 +0000 UTC m=+1018.244230332" Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.491600 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-87wk4"] Nov 25 10:53:41 crc kubenswrapper[4696]: I1125 10:53:41.498448 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-87wk4"] Nov 25 10:53:42 crc kubenswrapper[4696]: I1125 10:53:42.055620 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb781fb7-f7dc-468e-b326-38f1e5909127" path="/var/lib/kubelet/pods/eb781fb7-f7dc-468e-b326-38f1e5909127/volumes" Nov 25 10:53:43 crc kubenswrapper[4696]: I1125 10:53:43.452700 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e0d794d3-7682-4248-affe-b53f4f34f264","Type":"ContainerStarted","Data":"faa49d745ba8e7ba8e280a6a5296b1e1d9792ab47bcec384dc6bc1e37394d0be"} Nov 25 10:53:43 crc kubenswrapper[4696]: I1125 10:53:43.453169 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e0d794d3-7682-4248-affe-b53f4f34f264","Type":"ContainerStarted","Data":"195c106041e50ba23e8cbdf2ac1ce370ba9f292d406d2077e1213b445e06766e"} Nov 25 10:53:43 crc kubenswrapper[4696]: I1125 10:53:43.453191 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 10:53:44 crc kubenswrapper[4696]: I1125 10:53:44.465217 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"eba40268-f540-4ac7-9ae4-14455c81395f","Type":"ContainerStarted","Data":"d4080693bfaddf342d1276b6756cd2ee0719b20ae8e7307440f2417ebe7d8c87"} Nov 25 10:53:44 crc kubenswrapper[4696]: I1125 10:53:44.465751 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 10:53:44 crc kubenswrapper[4696]: I1125 10:53:44.492141 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.130919911 podStartE2EDuration="5.492120913s" podCreationTimestamp="2025-11-25 10:53:39 +0000 UTC" firstStartedPulling="2025-11-25 10:53:40.583053507 +0000 UTC m=+1017.396670094" lastFinishedPulling="2025-11-25 10:53:42.944254509 +0000 UTC m=+1019.757871096" observedRunningTime="2025-11-25 10:53:43.471491766 +0000 UTC m=+1020.285108393" watchObservedRunningTime="2025-11-25 10:53:44.492120913 +0000 UTC m=+1021.305737500" Nov 25 10:53:44 crc kubenswrapper[4696]: I1125 10:53:44.493479 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=6.674521089 podStartE2EDuration="46.493468045s" podCreationTimestamp="2025-11-25 10:52:58 +0000 UTC" firstStartedPulling="2025-11-25 10:53:03.619154587 +0000 UTC m=+980.432771174" lastFinishedPulling="2025-11-25 10:53:43.438101543 +0000 UTC m=+1020.251718130" observedRunningTime="2025-11-25 10:53:44.486620851 +0000 UTC m=+1021.300237468" watchObservedRunningTime="2025-11-25 10:53:44.493468045 +0000 UTC m=+1021.307084632" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.384024 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-924c-account-create-t4bcj"] Nov 25 10:53:48 crc kubenswrapper[4696]: E1125 10:53:48.384888 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb781fb7-f7dc-468e-b326-38f1e5909127" containerName="init" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.384902 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb781fb7-f7dc-468e-b326-38f1e5909127" containerName="init" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.385050 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb781fb7-f7dc-468e-b326-38f1e5909127" containerName="init" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.385508 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.392639 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.415830 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-924c-account-create-t4bcj"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.427835 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-xb2zp"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.429086 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.446022 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xb2zp"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.499309 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wt8l\" (UniqueName: \"kubernetes.io/projected/2815f305-2a4d-456e-af2c-0c842eda6acf-kube-api-access-6wt8l\") pod \"keystone-db-create-xb2zp\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.499370 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhh25\" (UniqueName: \"kubernetes.io/projected/7631c79c-d022-41b6-916e-102fcc294874-kube-api-access-hhh25\") pod \"keystone-924c-account-create-t4bcj\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.499391 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7631c79c-d022-41b6-916e-102fcc294874-operator-scripts\") pod \"keystone-924c-account-create-t4bcj\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.499442 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2815f305-2a4d-456e-af2c-0c842eda6acf-operator-scripts\") pod \"keystone-db-create-xb2zp\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.601243 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wt8l\" (UniqueName: \"kubernetes.io/projected/2815f305-2a4d-456e-af2c-0c842eda6acf-kube-api-access-6wt8l\") pod \"keystone-db-create-xb2zp\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.601351 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhh25\" (UniqueName: \"kubernetes.io/projected/7631c79c-d022-41b6-916e-102fcc294874-kube-api-access-hhh25\") pod \"keystone-924c-account-create-t4bcj\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.601391 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7631c79c-d022-41b6-916e-102fcc294874-operator-scripts\") pod \"keystone-924c-account-create-t4bcj\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.601485 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2815f305-2a4d-456e-af2c-0c842eda6acf-operator-scripts\") pod \"keystone-db-create-xb2zp\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.602780 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2815f305-2a4d-456e-af2c-0c842eda6acf-operator-scripts\") pod \"keystone-db-create-xb2zp\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.603136 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7631c79c-d022-41b6-916e-102fcc294874-operator-scripts\") pod \"keystone-924c-account-create-t4bcj\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.641565 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-d9gr4"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.643414 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.649472 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhh25\" (UniqueName: \"kubernetes.io/projected/7631c79c-d022-41b6-916e-102fcc294874-kube-api-access-hhh25\") pod \"keystone-924c-account-create-t4bcj\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.660159 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d9gr4"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.666972 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wt8l\" (UniqueName: \"kubernetes.io/projected/2815f305-2a4d-456e-af2c-0c842eda6acf-kube-api-access-6wt8l\") pod \"keystone-db-create-xb2zp\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.713019 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.726927 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmwk6\" (UniqueName: \"kubernetes.io/projected/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-kube-api-access-kmwk6\") pod \"placement-db-create-d9gr4\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.727291 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-operator-scripts\") pod \"placement-db-create-d9gr4\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.730820 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.753355 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.805517 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5ebe-account-create-7ffls"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.843857 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.852594 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.853990 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmwk6\" (UniqueName: \"kubernetes.io/projected/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-kube-api-access-kmwk6\") pod \"placement-db-create-d9gr4\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.854125 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-operator-scripts\") pod \"placement-db-create-d9gr4\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.855050 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-operator-scripts\") pod \"placement-db-create-d9gr4\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.871450 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5ebe-account-create-7ffls"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.880477 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmwk6\" (UniqueName: \"kubernetes.io/projected/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-kube-api-access-kmwk6\") pod \"placement-db-create-d9gr4\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.959196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtmlt\" (UniqueName: \"kubernetes.io/projected/ca998c2e-0f79-4bbe-8095-897dc135007c-kube-api-access-jtmlt\") pod \"placement-5ebe-account-create-7ffls\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.959295 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca998c2e-0f79-4bbe-8095-897dc135007c-operator-scripts\") pod \"placement-5ebe-account-create-7ffls\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.966943 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-p47td"] Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.968637 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-p47td" Nov 25 10:53:48 crc kubenswrapper[4696]: I1125 10:53:48.983500 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-p47td"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.062584 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtmlt\" (UniqueName: \"kubernetes.io/projected/ca998c2e-0f79-4bbe-8095-897dc135007c-kube-api-access-jtmlt\") pod \"placement-5ebe-account-create-7ffls\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.062680 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca998c2e-0f79-4bbe-8095-897dc135007c-operator-scripts\") pod \"placement-5ebe-account-create-7ffls\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.064246 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca998c2e-0f79-4bbe-8095-897dc135007c-operator-scripts\") pod \"placement-5ebe-account-create-7ffls\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.071582 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-63be-account-create-jxwx5"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.072550 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.078109 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.082595 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtmlt\" (UniqueName: \"kubernetes.io/projected/ca998c2e-0f79-4bbe-8095-897dc135007c-kube-api-access-jtmlt\") pod \"placement-5ebe-account-create-7ffls\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.083570 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-63be-account-create-jxwx5"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.164673 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-operator-scripts\") pod \"glance-db-create-p47td\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.164741 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkffz\" (UniqueName: \"kubernetes.io/projected/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-kube-api-access-dkffz\") pod \"glance-db-create-p47td\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.172944 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.195149 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.266230 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-operator-scripts\") pod \"glance-db-create-p47td\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.266601 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkffz\" (UniqueName: \"kubernetes.io/projected/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-kube-api-access-dkffz\") pod \"glance-db-create-p47td\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.266697 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f42bd\" (UniqueName: \"kubernetes.io/projected/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-kube-api-access-f42bd\") pod \"glance-63be-account-create-jxwx5\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.266739 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-operator-scripts\") pod \"glance-63be-account-create-jxwx5\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.268229 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-operator-scripts\") pod \"glance-db-create-p47td\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.294992 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkffz\" (UniqueName: \"kubernetes.io/projected/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-kube-api-access-dkffz\") pod \"glance-db-create-p47td\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.304864 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-p47td" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.359572 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xb2zp"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.365764 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-924c-account-create-t4bcj"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.368974 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f42bd\" (UniqueName: \"kubernetes.io/projected/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-kube-api-access-f42bd\") pod \"glance-63be-account-create-jxwx5\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.369020 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-operator-scripts\") pod \"glance-63be-account-create-jxwx5\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.376472 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-operator-scripts\") pod \"glance-63be-account-create-jxwx5\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: W1125 10:53:49.388063 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7631c79c_d022_41b6_916e_102fcc294874.slice/crio-f82ad9c5496265fd23a3cb9d24542b30cfd63270e19252493da07efa67bcc10d WatchSource:0}: Error finding container f82ad9c5496265fd23a3cb9d24542b30cfd63270e19252493da07efa67bcc10d: Status 404 returned error can't find the container with id f82ad9c5496265fd23a3cb9d24542b30cfd63270e19252493da07efa67bcc10d Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.396008 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f42bd\" (UniqueName: \"kubernetes.io/projected/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-kube-api-access-f42bd\") pod \"glance-63be-account-create-jxwx5\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.401035 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.490350 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d9gr4"] Nov 25 10:53:49 crc kubenswrapper[4696]: W1125 10:53:49.504178 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e192cbe_f2c6_44c0_ba41_81bf3e14705d.slice/crio-ef652bc80315c60df72e0889b39c81229cfbac9450e4129cf4be8534192382cb WatchSource:0}: Error finding container ef652bc80315c60df72e0889b39c81229cfbac9450e4129cf4be8534192382cb: Status 404 returned error can't find the container with id ef652bc80315c60df72e0889b39c81229cfbac9450e4129cf4be8534192382cb Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.526387 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-924c-account-create-t4bcj" event={"ID":"7631c79c-d022-41b6-916e-102fcc294874","Type":"ContainerStarted","Data":"f82ad9c5496265fd23a3cb9d24542b30cfd63270e19252493da07efa67bcc10d"} Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.527550 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xb2zp" event={"ID":"2815f305-2a4d-456e-af2c-0c842eda6acf","Type":"ContainerStarted","Data":"4506b4f49b142e0a8d6b79bb16a43405bcd3e223b6ae3a76a088a4467665fc17"} Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.546813 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.622497 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2mxjc"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.623343 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="dnsmasq-dns" containerID="cri-o://88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532" gracePeriod=10 Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.788175 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5ebe-account-create-7ffls"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.826296 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.99:5353: connect: connection refused" Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.927746 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-p47td"] Nov 25 10:53:49 crc kubenswrapper[4696]: I1125 10:53:49.992066 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-63be-account-create-jxwx5"] Nov 25 10:53:50 crc kubenswrapper[4696]: W1125 10:53:50.004747 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00abba76_7f70_4cb6_a4fa_46a44e2ebaac.slice/crio-ed3e79bcbc8062c80a147e7548f8ebf2a13e48a0a5d4321115e3b0223c113419 WatchSource:0}: Error finding container ed3e79bcbc8062c80a147e7548f8ebf2a13e48a0a5d4321115e3b0223c113419: Status 404 returned error can't find the container with id ed3e79bcbc8062c80a147e7548f8ebf2a13e48a0a5d4321115e3b0223c113419 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.202421 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.386162 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-config\") pod \"b5eaaf49-99f9-446b-a16d-aa792f882507\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.386269 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-dns-svc\") pod \"b5eaaf49-99f9-446b-a16d-aa792f882507\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.387257 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c987l\" (UniqueName: \"kubernetes.io/projected/b5eaaf49-99f9-446b-a16d-aa792f882507-kube-api-access-c987l\") pod \"b5eaaf49-99f9-446b-a16d-aa792f882507\" (UID: \"b5eaaf49-99f9-446b-a16d-aa792f882507\") " Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.392060 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5eaaf49-99f9-446b-a16d-aa792f882507-kube-api-access-c987l" (OuterVolumeSpecName: "kube-api-access-c987l") pod "b5eaaf49-99f9-446b-a16d-aa792f882507" (UID: "b5eaaf49-99f9-446b-a16d-aa792f882507"). InnerVolumeSpecName "kube-api-access-c987l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.427315 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-config" (OuterVolumeSpecName: "config") pod "b5eaaf49-99f9-446b-a16d-aa792f882507" (UID: "b5eaaf49-99f9-446b-a16d-aa792f882507"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.429326 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5eaaf49-99f9-446b-a16d-aa792f882507" (UID: "b5eaaf49-99f9-446b-a16d-aa792f882507"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.490081 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.490122 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5eaaf49-99f9-446b-a16d-aa792f882507-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.490138 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c987l\" (UniqueName: \"kubernetes.io/projected/b5eaaf49-99f9-446b-a16d-aa792f882507-kube-api-access-c987l\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.539757 4696 generic.go:334] "Generic (PLEG): container finished" podID="2815f305-2a4d-456e-af2c-0c842eda6acf" containerID="a8947ee445d01d5469d2336c922f90c9c89ec1d50ff83b761cbf2a20cabcfef4" exitCode=0 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.539834 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xb2zp" event={"ID":"2815f305-2a4d-456e-af2c-0c842eda6acf","Type":"ContainerDied","Data":"a8947ee445d01d5469d2336c922f90c9c89ec1d50ff83b761cbf2a20cabcfef4"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.556191 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-63be-account-create-jxwx5" event={"ID":"00abba76-7f70-4cb6-a4fa-46a44e2ebaac","Type":"ContainerStarted","Data":"257f611c064b71ea589c01025446c79f05ee6e3c5c6279ce562e5b2efbca806d"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.556249 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-63be-account-create-jxwx5" event={"ID":"00abba76-7f70-4cb6-a4fa-46a44e2ebaac","Type":"ContainerStarted","Data":"ed3e79bcbc8062c80a147e7548f8ebf2a13e48a0a5d4321115e3b0223c113419"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.559899 4696 generic.go:334] "Generic (PLEG): container finished" podID="7631c79c-d022-41b6-916e-102fcc294874" containerID="574ac7dfca7b3f18af5d2c4c4886c1374979b15a368d42c096248a8c7c1607da" exitCode=0 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.559962 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-924c-account-create-t4bcj" event={"ID":"7631c79c-d022-41b6-916e-102fcc294874","Type":"ContainerDied","Data":"574ac7dfca7b3f18af5d2c4c4886c1374979b15a368d42c096248a8c7c1607da"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.562008 4696 generic.go:334] "Generic (PLEG): container finished" podID="c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" containerID="fddf84327a2616525b0c6a26d45acab7c9e3acb6b06562585e089f24a871fff9" exitCode=0 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.562052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-p47td" event={"ID":"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34","Type":"ContainerDied","Data":"fddf84327a2616525b0c6a26d45acab7c9e3acb6b06562585e089f24a871fff9"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.562067 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-p47td" event={"ID":"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34","Type":"ContainerStarted","Data":"d5d22e17d6fcbb73be454f1b502db673f49e8254a92ac84e922da8e93f57cd65"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.570884 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e192cbe-f2c6-44c0-ba41-81bf3e14705d" containerID="f0d69789e1b7a66149e709d8a9c9e9c5497bb3d8ca3dbe145dacdde09fb24660" exitCode=0 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.570974 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d9gr4" event={"ID":"1e192cbe-f2c6-44c0-ba41-81bf3e14705d","Type":"ContainerDied","Data":"f0d69789e1b7a66149e709d8a9c9e9c5497bb3d8ca3dbe145dacdde09fb24660"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.570996 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d9gr4" event={"ID":"1e192cbe-f2c6-44c0-ba41-81bf3e14705d","Type":"ContainerStarted","Data":"ef652bc80315c60df72e0889b39c81229cfbac9450e4129cf4be8534192382cb"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.581284 4696 generic.go:334] "Generic (PLEG): container finished" podID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerID="88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532" exitCode=0 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.581885 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.582619 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" event={"ID":"b5eaaf49-99f9-446b-a16d-aa792f882507","Type":"ContainerDied","Data":"88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.582766 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2mxjc" event={"ID":"b5eaaf49-99f9-446b-a16d-aa792f882507","Type":"ContainerDied","Data":"6d0255020bc24c67baca1744e708f38b079a1c6d34f9b5935f52068b86af0baf"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.582788 4696 scope.go:117] "RemoveContainer" containerID="88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.585885 4696 generic.go:334] "Generic (PLEG): container finished" podID="ca998c2e-0f79-4bbe-8095-897dc135007c" containerID="1d45ffde46f1f111f12acbf0772017a3612d4ca37705c85102446cf1a0c8faf8" exitCode=0 Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.585939 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ebe-account-create-7ffls" event={"ID":"ca998c2e-0f79-4bbe-8095-897dc135007c","Type":"ContainerDied","Data":"1d45ffde46f1f111f12acbf0772017a3612d4ca37705c85102446cf1a0c8faf8"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.585967 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ebe-account-create-7ffls" event={"ID":"ca998c2e-0f79-4bbe-8095-897dc135007c","Type":"ContainerStarted","Data":"9e0627c81edf214c942b266357c2278c6979886d49a64cc8d024dea7a99fb3e7"} Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.624757 4696 scope.go:117] "RemoveContainer" containerID="a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.753233 4696 scope.go:117] "RemoveContainer" containerID="88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532" Nov 25 10:53:50 crc kubenswrapper[4696]: E1125 10:53:50.760181 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532\": container with ID starting with 88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532 not found: ID does not exist" containerID="88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.760224 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532"} err="failed to get container status \"88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532\": rpc error: code = NotFound desc = could not find container \"88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532\": container with ID starting with 88435740502e4006da722b96dda950a85f95940b2e1fa0cb861316bbba1af532 not found: ID does not exist" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.760249 4696 scope.go:117] "RemoveContainer" containerID="a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.767015 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2mxjc"] Nov 25 10:53:50 crc kubenswrapper[4696]: E1125 10:53:50.768644 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c\": container with ID starting with a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c not found: ID does not exist" containerID="a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.768687 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c"} err="failed to get container status \"a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c\": rpc error: code = NotFound desc = could not find container \"a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c\": container with ID starting with a093680f2141ed726572e1c7a250203a2513bcc797119f7fe832dbe1beb7433c not found: ID does not exist" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.771564 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2mxjc"] Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.833741 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-ljbcg"] Nov 25 10:53:50 crc kubenswrapper[4696]: E1125 10:53:50.834173 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="dnsmasq-dns" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.834199 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="dnsmasq-dns" Nov 25 10:53:50 crc kubenswrapper[4696]: E1125 10:53:50.834231 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="init" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.834242 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="init" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.834448 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" containerName="dnsmasq-dns" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.843568 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.876285 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-ljbcg"] Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.910047 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.910092 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.910130 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-config\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.910196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbvfq\" (UniqueName: \"kubernetes.io/projected/cbc24b3a-9e3a-4653-8617-45eac69ac198-kube-api-access-nbvfq\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:50 crc kubenswrapper[4696]: I1125 10:53:50.910215 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.012056 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbvfq\" (UniqueName: \"kubernetes.io/projected/cbc24b3a-9e3a-4653-8617-45eac69ac198-kube-api-access-nbvfq\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.012108 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.012155 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.012178 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.012220 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-config\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.013230 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.013257 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-config\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.013352 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.014067 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.050865 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbvfq\" (UniqueName: \"kubernetes.io/projected/cbc24b3a-9e3a-4653-8617-45eac69ac198-kube-api-access-nbvfq\") pod \"dnsmasq-dns-b8fbc5445-ljbcg\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.173008 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.597250 4696 generic.go:334] "Generic (PLEG): container finished" podID="00abba76-7f70-4cb6-a4fa-46a44e2ebaac" containerID="257f611c064b71ea589c01025446c79f05ee6e3c5c6279ce562e5b2efbca806d" exitCode=0 Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.597316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-63be-account-create-jxwx5" event={"ID":"00abba76-7f70-4cb6-a4fa-46a44e2ebaac","Type":"ContainerDied","Data":"257f611c064b71ea589c01025446c79f05ee6e3c5c6279ce562e5b2efbca806d"} Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.619210 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-ljbcg"] Nov 25 10:53:51 crc kubenswrapper[4696]: W1125 10:53:51.646571 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbc24b3a_9e3a_4653_8617_45eac69ac198.slice/crio-1a3e7aa58f78a4235688f6d6cf689c1ce1269b12dde90d0d46948c3b0cdee416 WatchSource:0}: Error finding container 1a3e7aa58f78a4235688f6d6cf689c1ce1269b12dde90d0d46948c3b0cdee416: Status 404 returned error can't find the container with id 1a3e7aa58f78a4235688f6d6cf689c1ce1269b12dde90d0d46948c3b0cdee416 Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.955632 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.968698 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.972228 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.972415 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.972525 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6mh7g" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.972642 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.976337 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 10:53:51 crc kubenswrapper[4696]: I1125 10:53:51.990595 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.034069 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.034214 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e87af5be-8499-4303-a158-8c9d50c2db73-lock\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.034414 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dss72\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-kube-api-access-dss72\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.034524 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.034584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e87af5be-8499-4303-a158-8c9d50c2db73-cache\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.058106 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5eaaf49-99f9-446b-a16d-aa792f882507" path="/var/lib/kubelet/pods/b5eaaf49-99f9-446b-a16d-aa792f882507/volumes" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.135635 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtmlt\" (UniqueName: \"kubernetes.io/projected/ca998c2e-0f79-4bbe-8095-897dc135007c-kube-api-access-jtmlt\") pod \"ca998c2e-0f79-4bbe-8095-897dc135007c\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.137208 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca998c2e-0f79-4bbe-8095-897dc135007c-operator-scripts\") pod \"ca998c2e-0f79-4bbe-8095-897dc135007c\" (UID: \"ca998c2e-0f79-4bbe-8095-897dc135007c\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.137618 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e87af5be-8499-4303-a158-8c9d50c2db73-cache\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.137770 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.138102 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e87af5be-8499-4303-a158-8c9d50c2db73-lock\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.138181 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dss72\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-kube-api-access-dss72\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.138240 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.138465 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.138480 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.138540 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift podName:e87af5be-8499-4303-a158-8c9d50c2db73 nodeName:}" failed. No retries permitted until 2025-11-25 10:53:52.638502838 +0000 UTC m=+1029.452119415 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift") pod "swift-storage-0" (UID: "e87af5be-8499-4303-a158-8c9d50c2db73") : configmap "swift-ring-files" not found Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.138605 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e87af5be-8499-4303-a158-8c9d50c2db73-cache\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.139453 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca998c2e-0f79-4bbe-8095-897dc135007c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ca998c2e-0f79-4bbe-8095-897dc135007c" (UID: "ca998c2e-0f79-4bbe-8095-897dc135007c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.140091 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e87af5be-8499-4303-a158-8c9d50c2db73-lock\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.140364 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.144432 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.146714 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca998c2e-0f79-4bbe-8095-897dc135007c-kube-api-access-jtmlt" (OuterVolumeSpecName: "kube-api-access-jtmlt") pod "ca998c2e-0f79-4bbe-8095-897dc135007c" (UID: "ca998c2e-0f79-4bbe-8095-897dc135007c"). InnerVolumeSpecName "kube-api-access-jtmlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.161442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dss72\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-kube-api-access-dss72\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.172332 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.189237 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.220463 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.245624 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtmlt\" (UniqueName: \"kubernetes.io/projected/ca998c2e-0f79-4bbe-8095-897dc135007c-kube-api-access-jtmlt\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.245677 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca998c2e-0f79-4bbe-8095-897dc135007c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.346244 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhh25\" (UniqueName: \"kubernetes.io/projected/7631c79c-d022-41b6-916e-102fcc294874-kube-api-access-hhh25\") pod \"7631c79c-d022-41b6-916e-102fcc294874\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.346290 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-operator-scripts\") pod \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.346376 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wt8l\" (UniqueName: \"kubernetes.io/projected/2815f305-2a4d-456e-af2c-0c842eda6acf-kube-api-access-6wt8l\") pod \"2815f305-2a4d-456e-af2c-0c842eda6acf\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.346401 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2815f305-2a4d-456e-af2c-0c842eda6acf-operator-scripts\") pod \"2815f305-2a4d-456e-af2c-0c842eda6acf\" (UID: \"2815f305-2a4d-456e-af2c-0c842eda6acf\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.346435 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f42bd\" (UniqueName: \"kubernetes.io/projected/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-kube-api-access-f42bd\") pod \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\" (UID: \"00abba76-7f70-4cb6-a4fa-46a44e2ebaac\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.346503 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7631c79c-d022-41b6-916e-102fcc294874-operator-scripts\") pod \"7631c79c-d022-41b6-916e-102fcc294874\" (UID: \"7631c79c-d022-41b6-916e-102fcc294874\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.347739 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2815f305-2a4d-456e-af2c-0c842eda6acf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2815f305-2a4d-456e-af2c-0c842eda6acf" (UID: "2815f305-2a4d-456e-af2c-0c842eda6acf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.348421 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7631c79c-d022-41b6-916e-102fcc294874-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7631c79c-d022-41b6-916e-102fcc294874" (UID: "7631c79c-d022-41b6-916e-102fcc294874"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.349464 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00abba76-7f70-4cb6-a4fa-46a44e2ebaac" (UID: "00abba76-7f70-4cb6-a4fa-46a44e2ebaac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.350389 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2815f305-2a4d-456e-af2c-0c842eda6acf-kube-api-access-6wt8l" (OuterVolumeSpecName: "kube-api-access-6wt8l") pod "2815f305-2a4d-456e-af2c-0c842eda6acf" (UID: "2815f305-2a4d-456e-af2c-0c842eda6acf"). InnerVolumeSpecName "kube-api-access-6wt8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.352077 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-kube-api-access-f42bd" (OuterVolumeSpecName: "kube-api-access-f42bd") pod "00abba76-7f70-4cb6-a4fa-46a44e2ebaac" (UID: "00abba76-7f70-4cb6-a4fa-46a44e2ebaac"). InnerVolumeSpecName "kube-api-access-f42bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.354959 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.363823 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7631c79c-d022-41b6-916e-102fcc294874-kube-api-access-hhh25" (OuterVolumeSpecName: "kube-api-access-hhh25") pod "7631c79c-d022-41b6-916e-102fcc294874" (UID: "7631c79c-d022-41b6-916e-102fcc294874"). InnerVolumeSpecName "kube-api-access-hhh25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.375931 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-p47td" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.452695 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wt8l\" (UniqueName: \"kubernetes.io/projected/2815f305-2a4d-456e-af2c-0c842eda6acf-kube-api-access-6wt8l\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.452724 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2815f305-2a4d-456e-af2c-0c842eda6acf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.452735 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f42bd\" (UniqueName: \"kubernetes.io/projected/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-kube-api-access-f42bd\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.452744 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7631c79c-d022-41b6-916e-102fcc294874-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.452753 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhh25\" (UniqueName: \"kubernetes.io/projected/7631c79c-d022-41b6-916e-102fcc294874-kube-api-access-hhh25\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.452764 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00abba76-7f70-4cb6-a4fa-46a44e2ebaac-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460002 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-xshf5"] Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.460355 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e192cbe-f2c6-44c0-ba41-81bf3e14705d" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460375 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e192cbe-f2c6-44c0-ba41-81bf3e14705d" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.460394 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460401 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.460413 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7631c79c-d022-41b6-916e-102fcc294874" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460420 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7631c79c-d022-41b6-916e-102fcc294874" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.460429 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2815f305-2a4d-456e-af2c-0c842eda6acf" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460435 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2815f305-2a4d-456e-af2c-0c842eda6acf" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.460446 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00abba76-7f70-4cb6-a4fa-46a44e2ebaac" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460451 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="00abba76-7f70-4cb6-a4fa-46a44e2ebaac" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.460462 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca998c2e-0f79-4bbe-8095-897dc135007c" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460467 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca998c2e-0f79-4bbe-8095-897dc135007c" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460631 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7631c79c-d022-41b6-916e-102fcc294874" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460639 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460650 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2815f305-2a4d-456e-af2c-0c842eda6acf" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460673 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca998c2e-0f79-4bbe-8095-897dc135007c" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460691 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="00abba76-7f70-4cb6-a4fa-46a44e2ebaac" containerName="mariadb-account-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.460701 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e192cbe-f2c6-44c0-ba41-81bf3e14705d" containerName="mariadb-database-create" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.466599 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.472346 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.472641 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.472930 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.489643 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xshf5"] Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.554013 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkffz\" (UniqueName: \"kubernetes.io/projected/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-kube-api-access-dkffz\") pod \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.554174 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmwk6\" (UniqueName: \"kubernetes.io/projected/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-kube-api-access-kmwk6\") pod \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.554230 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-operator-scripts\") pod \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\" (UID: \"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.554254 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-operator-scripts\") pod \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\" (UID: \"1e192cbe-f2c6-44c0-ba41-81bf3e14705d\") " Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.555016 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" (UID: "c5ff86ec-50a8-4137-8f0e-b1adb75ffe34"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.555085 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e192cbe-f2c6-44c0-ba41-81bf3e14705d" (UID: "1e192cbe-f2c6-44c0-ba41-81bf3e14705d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.557786 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-kube-api-access-dkffz" (OuterVolumeSpecName: "kube-api-access-dkffz") pod "c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" (UID: "c5ff86ec-50a8-4137-8f0e-b1adb75ffe34"). InnerVolumeSpecName "kube-api-access-dkffz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.559758 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-kube-api-access-kmwk6" (OuterVolumeSpecName: "kube-api-access-kmwk6") pod "1e192cbe-f2c6-44c0-ba41-81bf3e14705d" (UID: "1e192cbe-f2c6-44c0-ba41-81bf3e14705d"). InnerVolumeSpecName "kube-api-access-kmwk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.611044 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-63be-account-create-jxwx5" event={"ID":"00abba76-7f70-4cb6-a4fa-46a44e2ebaac","Type":"ContainerDied","Data":"ed3e79bcbc8062c80a147e7548f8ebf2a13e48a0a5d4321115e3b0223c113419"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.611746 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed3e79bcbc8062c80a147e7548f8ebf2a13e48a0a5d4321115e3b0223c113419" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.611851 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-63be-account-create-jxwx5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.614875 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-924c-account-create-t4bcj" event={"ID":"7631c79c-d022-41b6-916e-102fcc294874","Type":"ContainerDied","Data":"f82ad9c5496265fd23a3cb9d24542b30cfd63270e19252493da07efa67bcc10d"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.614921 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f82ad9c5496265fd23a3cb9d24542b30cfd63270e19252493da07efa67bcc10d" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.614924 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-924c-account-create-t4bcj" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.615882 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-p47td" event={"ID":"c5ff86ec-50a8-4137-8f0e-b1adb75ffe34","Type":"ContainerDied","Data":"d5d22e17d6fcbb73be454f1b502db673f49e8254a92ac84e922da8e93f57cd65"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.615977 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5d22e17d6fcbb73be454f1b502db673f49e8254a92ac84e922da8e93f57cd65" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.616092 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-p47td" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.619725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d9gr4" event={"ID":"1e192cbe-f2c6-44c0-ba41-81bf3e14705d","Type":"ContainerDied","Data":"ef652bc80315c60df72e0889b39c81229cfbac9450e4129cf4be8534192382cb"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.619761 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef652bc80315c60df72e0889b39c81229cfbac9450e4129cf4be8534192382cb" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.619765 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d9gr4" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.622265 4696 generic.go:334] "Generic (PLEG): container finished" podID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerID="50f861584f6abe0a40b4b1df28ec8c6f39277768349ee3db6222df4a73805250" exitCode=0 Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.622321 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" event={"ID":"cbc24b3a-9e3a-4653-8617-45eac69ac198","Type":"ContainerDied","Data":"50f861584f6abe0a40b4b1df28ec8c6f39277768349ee3db6222df4a73805250"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.622347 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" event={"ID":"cbc24b3a-9e3a-4653-8617-45eac69ac198","Type":"ContainerStarted","Data":"1a3e7aa58f78a4235688f6d6cf689c1ce1269b12dde90d0d46948c3b0cdee416"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.628647 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ebe-account-create-7ffls" event={"ID":"ca998c2e-0f79-4bbe-8095-897dc135007c","Type":"ContainerDied","Data":"9e0627c81edf214c942b266357c2278c6979886d49a64cc8d024dea7a99fb3e7"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.628702 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e0627c81edf214c942b266357c2278c6979886d49a64cc8d024dea7a99fb3e7" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.628790 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ebe-account-create-7ffls" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.632415 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xb2zp" event={"ID":"2815f305-2a4d-456e-af2c-0c842eda6acf","Type":"ContainerDied","Data":"4506b4f49b142e0a8d6b79bb16a43405bcd3e223b6ae3a76a088a4467665fc17"} Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.632443 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4506b4f49b142e0a8d6b79bb16a43405bcd3e223b6ae3a76a088a4467665fc17" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.632560 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xb2zp" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.655950 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-scripts\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.656856 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-ring-data-devices\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.656875 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-dispersionconf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.656932 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-swiftconf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.656981 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657006 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-combined-ca-bundle\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657024 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzqkf\" (UniqueName: \"kubernetes.io/projected/dbeda82d-3956-47f7-95a8-6ffc25961e5d-kube-api-access-nzqkf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657045 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dbeda82d-3956-47f7-95a8-6ffc25961e5d-etc-swift\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657088 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmwk6\" (UniqueName: \"kubernetes.io/projected/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-kube-api-access-kmwk6\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657099 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657108 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e192cbe-f2c6-44c0-ba41-81bf3e14705d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.657116 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkffz\" (UniqueName: \"kubernetes.io/projected/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34-kube-api-access-dkffz\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.657303 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.657908 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:53:52 crc kubenswrapper[4696]: E1125 10:53:52.657951 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift podName:e87af5be-8499-4303-a158-8c9d50c2db73 nodeName:}" failed. No retries permitted until 2025-11-25 10:53:53.657937501 +0000 UTC m=+1030.471554088 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift") pod "swift-storage-0" (UID: "e87af5be-8499-4303-a158-8c9d50c2db73") : configmap "swift-ring-files" not found Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.758925 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-swiftconf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.759035 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-combined-ca-bundle\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.759059 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzqkf\" (UniqueName: \"kubernetes.io/projected/dbeda82d-3956-47f7-95a8-6ffc25961e5d-kube-api-access-nzqkf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.759079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dbeda82d-3956-47f7-95a8-6ffc25961e5d-etc-swift\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.759123 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-scripts\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.759896 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-ring-data-devices\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.759170 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-ring-data-devices\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.765618 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-dispersionconf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.768583 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-swiftconf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.769039 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dbeda82d-3956-47f7-95a8-6ffc25961e5d-etc-swift\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.772336 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-dispersionconf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.772999 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-scripts\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.783899 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-combined-ca-bundle\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:52 crc kubenswrapper[4696]: I1125 10:53:52.791223 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzqkf\" (UniqueName: \"kubernetes.io/projected/dbeda82d-3956-47f7-95a8-6ffc25961e5d-kube-api-access-nzqkf\") pod \"swift-ring-rebalance-xshf5\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.089944 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.568831 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xshf5"] Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.640427 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" event={"ID":"cbc24b3a-9e3a-4653-8617-45eac69ac198","Type":"ContainerStarted","Data":"a196b7680557afe9147c63a12d5a58855670014218b4c017d107fdab61427c25"} Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.641350 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.642969 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xshf5" event={"ID":"dbeda82d-3956-47f7-95a8-6ffc25961e5d","Type":"ContainerStarted","Data":"cce5513a48e7aab4fda1bad9c488483ad6c7688721d0c47b5358bafe9f8592c0"} Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.677372 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" podStartSLOduration=3.6773531 podStartE2EDuration="3.6773531s" podCreationTimestamp="2025-11-25 10:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:53.671689723 +0000 UTC m=+1030.485306310" watchObservedRunningTime="2025-11-25 10:53:53.6773531 +0000 UTC m=+1030.490969677" Nov 25 10:53:53 crc kubenswrapper[4696]: E1125 10:53:53.685677 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:53:53 crc kubenswrapper[4696]: E1125 10:53:53.685703 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:53:53 crc kubenswrapper[4696]: I1125 10:53:53.686045 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:53 crc kubenswrapper[4696]: E1125 10:53:53.686104 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift podName:e87af5be-8499-4303-a158-8c9d50c2db73 nodeName:}" failed. No retries permitted until 2025-11-25 10:53:55.686085603 +0000 UTC m=+1032.499702180 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift") pod "swift-storage-0" (UID: "e87af5be-8499-4303-a158-8c9d50c2db73") : configmap "swift-ring-files" not found Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.191127 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ljbtx"] Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.192169 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.197401 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-s5c2l" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.197643 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.214698 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ljbtx"] Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.294508 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-config-data\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.294884 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqkgx\" (UniqueName: \"kubernetes.io/projected/bc2593ea-7985-4050-8000-2be38f4b488c-kube-api-access-qqkgx\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.294916 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-db-sync-config-data\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.294944 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-combined-ca-bundle\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.396020 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqkgx\" (UniqueName: \"kubernetes.io/projected/bc2593ea-7985-4050-8000-2be38f4b488c-kube-api-access-qqkgx\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.396124 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-db-sync-config-data\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.397155 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-combined-ca-bundle\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.397325 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-config-data\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.412001 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-config-data\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.412449 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-combined-ca-bundle\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.414393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-db-sync-config-data\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.414921 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqkgx\" (UniqueName: \"kubernetes.io/projected/bc2593ea-7985-4050-8000-2be38f4b488c-kube-api-access-qqkgx\") pod \"glance-db-sync-ljbtx\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:54 crc kubenswrapper[4696]: I1125 10:53:54.520164 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ljbtx" Nov 25 10:53:55 crc kubenswrapper[4696]: I1125 10:53:55.054912 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ljbtx"] Nov 25 10:53:55 crc kubenswrapper[4696]: I1125 10:53:55.299950 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 10:53:55 crc kubenswrapper[4696]: I1125 10:53:55.666657 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ljbtx" event={"ID":"bc2593ea-7985-4050-8000-2be38f4b488c","Type":"ContainerStarted","Data":"b9823b1a9f41cb38b25426e3eda10ea6bd4d5940999cf67a2e6e0fdcba4d79ae"} Nov 25 10:53:55 crc kubenswrapper[4696]: I1125 10:53:55.730067 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:55 crc kubenswrapper[4696]: E1125 10:53:55.730303 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:53:55 crc kubenswrapper[4696]: E1125 10:53:55.730318 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:53:55 crc kubenswrapper[4696]: E1125 10:53:55.730359 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift podName:e87af5be-8499-4303-a158-8c9d50c2db73 nodeName:}" failed. No retries permitted until 2025-11-25 10:53:59.730346241 +0000 UTC m=+1036.543962828 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift") pod "swift-storage-0" (UID: "e87af5be-8499-4303-a158-8c9d50c2db73") : configmap "swift-ring-files" not found Nov 25 10:53:57 crc kubenswrapper[4696]: I1125 10:53:57.686211 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xshf5" event={"ID":"dbeda82d-3956-47f7-95a8-6ffc25961e5d","Type":"ContainerStarted","Data":"00f6ef218ecad77c6b0c2bacac6a6fecbf122472a39d4a7cd72c533984615290"} Nov 25 10:53:57 crc kubenswrapper[4696]: I1125 10:53:57.711914 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-xshf5" podStartSLOduration=2.137652456 podStartE2EDuration="5.711897598s" podCreationTimestamp="2025-11-25 10:53:52 +0000 UTC" firstStartedPulling="2025-11-25 10:53:53.575634202 +0000 UTC m=+1030.389250789" lastFinishedPulling="2025-11-25 10:53:57.149879334 +0000 UTC m=+1033.963495931" observedRunningTime="2025-11-25 10:53:57.706468779 +0000 UTC m=+1034.520085376" watchObservedRunningTime="2025-11-25 10:53:57.711897598 +0000 UTC m=+1034.525514185" Nov 25 10:53:59 crc kubenswrapper[4696]: I1125 10:53:59.707175 4696 generic.go:334] "Generic (PLEG): container finished" podID="01da0ce5-e289-4f9d-95d2-922e92053827" containerID="c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b" exitCode=0 Nov 25 10:53:59 crc kubenswrapper[4696]: I1125 10:53:59.707282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01da0ce5-e289-4f9d-95d2-922e92053827","Type":"ContainerDied","Data":"c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b"} Nov 25 10:53:59 crc kubenswrapper[4696]: I1125 10:53:59.805526 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:53:59 crc kubenswrapper[4696]: E1125 10:53:59.806191 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:53:59 crc kubenswrapper[4696]: E1125 10:53:59.806214 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:53:59 crc kubenswrapper[4696]: E1125 10:53:59.806254 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift podName:e87af5be-8499-4303-a158-8c9d50c2db73 nodeName:}" failed. No retries permitted until 2025-11-25 10:54:07.806238021 +0000 UTC m=+1044.619854598 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift") pod "swift-storage-0" (UID: "e87af5be-8499-4303-a158-8c9d50c2db73") : configmap "swift-ring-files" not found Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.732555 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01da0ce5-e289-4f9d-95d2-922e92053827","Type":"ContainerStarted","Data":"9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6"} Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.734021 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.737115 4696 generic.go:334] "Generic (PLEG): container finished" podID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerID="fa9cbf49307a6d897269a72edeb254cafcbb7dab2a48856998a38a4b27815fba" exitCode=0 Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.737153 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2ff36659-0a3a-4a6f-9f79-414918df363a","Type":"ContainerDied","Data":"fa9cbf49307a6d897269a72edeb254cafcbb7dab2a48856998a38a4b27815fba"} Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.781543 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371970.073256 podStartE2EDuration="1m6.781521181s" podCreationTimestamp="2025-11-25 10:52:54 +0000 UTC" firstStartedPulling="2025-11-25 10:52:56.582733222 +0000 UTC m=+973.396349809" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:00.760147262 +0000 UTC m=+1037.573763869" watchObservedRunningTime="2025-11-25 10:54:00.781521181 +0000 UTC m=+1037.595137768" Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.801419 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:54:00 crc kubenswrapper[4696]: I1125 10:54:00.801465 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.174882 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.233393 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tppmp"] Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.247821 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-tppmp" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerName="dnsmasq-dns" containerID="cri-o://08d01a7c885b90c8624821d4a7447527f42bfcd77f3fddba9bbb010fa6bce2e3" gracePeriod=10 Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.751734 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2ff36659-0a3a-4a6f-9f79-414918df363a","Type":"ContainerStarted","Data":"6b68ac85afb11c275db7068ddade9c64bd80c7ef050364e5e752636e7879b7db"} Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.752409 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.755982 4696 generic.go:334] "Generic (PLEG): container finished" podID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerID="08d01a7c885b90c8624821d4a7447527f42bfcd77f3fddba9bbb010fa6bce2e3" exitCode=0 Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.756019 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tppmp" event={"ID":"a67dc117-d4d1-4cf3-bd47-0683bfc7b108","Type":"ContainerDied","Data":"08d01a7c885b90c8624821d4a7447527f42bfcd77f3fddba9bbb010fa6bce2e3"} Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.778828 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.026463256 podStartE2EDuration="1m7.778806808s" podCreationTimestamp="2025-11-25 10:52:54 +0000 UTC" firstStartedPulling="2025-11-25 10:52:56.699854508 +0000 UTC m=+973.513471095" lastFinishedPulling="2025-11-25 10:53:27.45219806 +0000 UTC m=+1004.265814647" observedRunningTime="2025-11-25 10:54:01.772062447 +0000 UTC m=+1038.585679034" watchObservedRunningTime="2025-11-25 10:54:01.778806808 +0000 UTC m=+1038.592423395" Nov 25 10:54:01 crc kubenswrapper[4696]: I1125 10:54:01.940444 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.051599 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-sb\") pod \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.051777 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-dns-svc\") pod \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.051906 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-nb\") pod \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.051982 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg88r\" (UniqueName: \"kubernetes.io/projected/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-kube-api-access-fg88r\") pod \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.052115 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-config\") pod \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\" (UID: \"a67dc117-d4d1-4cf3-bd47-0683bfc7b108\") " Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.058629 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-kube-api-access-fg88r" (OuterVolumeSpecName: "kube-api-access-fg88r") pod "a67dc117-d4d1-4cf3-bd47-0683bfc7b108" (UID: "a67dc117-d4d1-4cf3-bd47-0683bfc7b108"). InnerVolumeSpecName "kube-api-access-fg88r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.150344 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a67dc117-d4d1-4cf3-bd47-0683bfc7b108" (UID: "a67dc117-d4d1-4cf3-bd47-0683bfc7b108"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.158993 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a67dc117-d4d1-4cf3-bd47-0683bfc7b108" (UID: "a67dc117-d4d1-4cf3-bd47-0683bfc7b108"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.159067 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.159082 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg88r\" (UniqueName: \"kubernetes.io/projected/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-kube-api-access-fg88r\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.171615 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a67dc117-d4d1-4cf3-bd47-0683bfc7b108" (UID: "a67dc117-d4d1-4cf3-bd47-0683bfc7b108"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.189651 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-config" (OuterVolumeSpecName: "config") pod "a67dc117-d4d1-4cf3-bd47-0683bfc7b108" (UID: "a67dc117-d4d1-4cf3-bd47-0683bfc7b108"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.260838 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.260871 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.260882 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67dc117-d4d1-4cf3-bd47-0683bfc7b108-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.782998 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tppmp" event={"ID":"a67dc117-d4d1-4cf3-bd47-0683bfc7b108","Type":"ContainerDied","Data":"0c108cf5f9793277ac0b45699a263c53b4beea4515b27b4a590fcf410f3fa8a6"} Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.783062 4696 scope.go:117] "RemoveContainer" containerID="08d01a7c885b90c8624821d4a7447527f42bfcd77f3fddba9bbb010fa6bce2e3" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.783241 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tppmp" Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.821653 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tppmp"] Nov 25 10:54:02 crc kubenswrapper[4696]: I1125 10:54:02.832638 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tppmp"] Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.055268 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" path="/var/lib/kubelet/pods/a67dc117-d4d1-4cf3-bd47-0683bfc7b108/volumes" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.533869 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jz4h7" podUID="87dc2400-74ac-4b0b-ae7d-ca62010572d3" containerName="ovn-controller" probeResult="failure" output=< Nov 25 10:54:04 crc kubenswrapper[4696]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 10:54:04 crc kubenswrapper[4696]: > Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.558789 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.572244 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7w2gw" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.794688 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jz4h7-config-z2gpm"] Nov 25 10:54:04 crc kubenswrapper[4696]: E1125 10:54:04.797959 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerName="dnsmasq-dns" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.798248 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerName="dnsmasq-dns" Nov 25 10:54:04 crc kubenswrapper[4696]: E1125 10:54:04.798322 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerName="init" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.798384 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerName="init" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.798657 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a67dc117-d4d1-4cf3-bd47-0683bfc7b108" containerName="dnsmasq-dns" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.799268 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.807018 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.810841 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jz4h7-config-z2gpm"] Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.822934 4696 generic.go:334] "Generic (PLEG): container finished" podID="dbeda82d-3956-47f7-95a8-6ffc25961e5d" containerID="00f6ef218ecad77c6b0c2bacac6a6fecbf122472a39d4a7cd72c533984615290" exitCode=0 Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.824282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xshf5" event={"ID":"dbeda82d-3956-47f7-95a8-6ffc25961e5d","Type":"ContainerDied","Data":"00f6ef218ecad77c6b0c2bacac6a6fecbf122472a39d4a7cd72c533984615290"} Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.899461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-additional-scripts\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.899828 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run-ovn\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.899928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.900005 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-log-ovn\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.900143 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-scripts\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:04 crc kubenswrapper[4696]: I1125 10:54:04.900250 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsgjp\" (UniqueName: \"kubernetes.io/projected/7d786afc-ebdc-4806-9daf-86aa8d2563dd-kube-api-access-gsgjp\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002208 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-scripts\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002278 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsgjp\" (UniqueName: \"kubernetes.io/projected/7d786afc-ebdc-4806-9daf-86aa8d2563dd-kube-api-access-gsgjp\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002310 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-additional-scripts\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002377 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run-ovn\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002408 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002427 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-log-ovn\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.002790 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-log-ovn\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.004728 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-scripts\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.005483 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-additional-scripts\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.005552 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run-ovn\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.005601 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.026316 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsgjp\" (UniqueName: \"kubernetes.io/projected/7d786afc-ebdc-4806-9daf-86aa8d2563dd-kube-api-access-gsgjp\") pod \"ovn-controller-jz4h7-config-z2gpm\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:05 crc kubenswrapper[4696]: I1125 10:54:05.117897 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:07 crc kubenswrapper[4696]: I1125 10:54:07.851120 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:54:07 crc kubenswrapper[4696]: I1125 10:54:07.857393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e87af5be-8499-4303-a158-8c9d50c2db73-etc-swift\") pod \"swift-storage-0\" (UID: \"e87af5be-8499-4303-a158-8c9d50c2db73\") " pod="openstack/swift-storage-0" Nov 25 10:54:07 crc kubenswrapper[4696]: I1125 10:54:07.925080 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 10:54:09 crc kubenswrapper[4696]: I1125 10:54:09.504272 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jz4h7" podUID="87dc2400-74ac-4b0b-ae7d-ca62010572d3" containerName="ovn-controller" probeResult="failure" output=< Nov 25 10:54:09 crc kubenswrapper[4696]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 10:54:09 crc kubenswrapper[4696]: > Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.424156 4696 scope.go:117] "RemoveContainer" containerID="9eff8cb02b324b2fd51b27e80202509f61a1cc0a23b3ea7340663eba2deff5d8" Nov 25 10:54:11 crc kubenswrapper[4696]: E1125 10:54:11.456695 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Nov 25 10:54:11 crc kubenswrapper[4696]: E1125 10:54:11.456843 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qqkgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-ljbtx_openstack(bc2593ea-7985-4050-8000-2be38f4b488c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:54:11 crc kubenswrapper[4696]: E1125 10:54:11.457992 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-ljbtx" podUID="bc2593ea-7985-4050-8000-2be38f4b488c" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.580936 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720173 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-swiftconf\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720215 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-ring-data-devices\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720280 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-combined-ca-bundle\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720320 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-scripts\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720384 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-dispersionconf\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720410 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dbeda82d-3956-47f7-95a8-6ffc25961e5d-etc-swift\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.720856 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzqkf\" (UniqueName: \"kubernetes.io/projected/dbeda82d-3956-47f7-95a8-6ffc25961e5d-kube-api-access-nzqkf\") pod \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\" (UID: \"dbeda82d-3956-47f7-95a8-6ffc25961e5d\") " Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.721269 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbeda82d-3956-47f7-95a8-6ffc25961e5d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.721649 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.759343 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.761894 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbeda82d-3956-47f7-95a8-6ffc25961e5d-kube-api-access-nzqkf" (OuterVolumeSpecName: "kube-api-access-nzqkf") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "kube-api-access-nzqkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.766912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.769950 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.796408 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-scripts" (OuterVolumeSpecName: "scripts") pod "dbeda82d-3956-47f7-95a8-6ffc25961e5d" (UID: "dbeda82d-3956-47f7-95a8-6ffc25961e5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.822973 4696 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.822998 4696 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.823008 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.823016 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbeda82d-3956-47f7-95a8-6ffc25961e5d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.823024 4696 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dbeda82d-3956-47f7-95a8-6ffc25961e5d-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.823031 4696 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dbeda82d-3956-47f7-95a8-6ffc25961e5d-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.823041 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzqkf\" (UniqueName: \"kubernetes.io/projected/dbeda82d-3956-47f7-95a8-6ffc25961e5d-kube-api-access-nzqkf\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.882564 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xshf5" Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.882869 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xshf5" event={"ID":"dbeda82d-3956-47f7-95a8-6ffc25961e5d","Type":"ContainerDied","Data":"cce5513a48e7aab4fda1bad9c488483ad6c7688721d0c47b5358bafe9f8592c0"} Nov 25 10:54:11 crc kubenswrapper[4696]: I1125 10:54:11.882914 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cce5513a48e7aab4fda1bad9c488483ad6c7688721d0c47b5358bafe9f8592c0" Nov 25 10:54:11 crc kubenswrapper[4696]: E1125 10:54:11.887025 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-ljbtx" podUID="bc2593ea-7985-4050-8000-2be38f4b488c" Nov 25 10:54:12 crc kubenswrapper[4696]: W1125 10:54:12.021654 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode87af5be_8499_4303_a158_8c9d50c2db73.slice/crio-0b1e88a2f6566afb6bc9b5df4fa16455a59ac8031f0fe171e6ca226883c65ae8 WatchSource:0}: Error finding container 0b1e88a2f6566afb6bc9b5df4fa16455a59ac8031f0fe171e6ca226883c65ae8: Status 404 returned error can't find the container with id 0b1e88a2f6566afb6bc9b5df4fa16455a59ac8031f0fe171e6ca226883c65ae8 Nov 25 10:54:12 crc kubenswrapper[4696]: I1125 10:54:12.029488 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 10:54:12 crc kubenswrapper[4696]: W1125 10:54:12.037206 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d786afc_ebdc_4806_9daf_86aa8d2563dd.slice/crio-1e347649ce5f0f16b39a637986ff5c3e6dbe356d7856e8220e6f3e86f99227e0 WatchSource:0}: Error finding container 1e347649ce5f0f16b39a637986ff5c3e6dbe356d7856e8220e6f3e86f99227e0: Status 404 returned error can't find the container with id 1e347649ce5f0f16b39a637986ff5c3e6dbe356d7856e8220e6f3e86f99227e0 Nov 25 10:54:12 crc kubenswrapper[4696]: I1125 10:54:12.038309 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jz4h7-config-z2gpm"] Nov 25 10:54:12 crc kubenswrapper[4696]: I1125 10:54:12.903192 4696 generic.go:334] "Generic (PLEG): container finished" podID="7d786afc-ebdc-4806-9daf-86aa8d2563dd" containerID="cdb45bee6c0bc35518f191752999d1266038fe12e4b900064b46879e341b322d" exitCode=0 Nov 25 10:54:12 crc kubenswrapper[4696]: I1125 10:54:12.903278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jz4h7-config-z2gpm" event={"ID":"7d786afc-ebdc-4806-9daf-86aa8d2563dd","Type":"ContainerDied","Data":"cdb45bee6c0bc35518f191752999d1266038fe12e4b900064b46879e341b322d"} Nov 25 10:54:12 crc kubenswrapper[4696]: I1125 10:54:12.903324 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jz4h7-config-z2gpm" event={"ID":"7d786afc-ebdc-4806-9daf-86aa8d2563dd","Type":"ContainerStarted","Data":"1e347649ce5f0f16b39a637986ff5c3e6dbe356d7856e8220e6f3e86f99227e0"} Nov 25 10:54:12 crc kubenswrapper[4696]: I1125 10:54:12.912756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"0b1e88a2f6566afb6bc9b5df4fa16455a59ac8031f0fe171e6ca226883c65ae8"} Nov 25 10:54:13 crc kubenswrapper[4696]: I1125 10:54:13.921153 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"003aae161034d06c283137573d5d8e87ed762eee65bf968d3317b95440aa48da"} Nov 25 10:54:13 crc kubenswrapper[4696]: I1125 10:54:13.921441 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"b43775704cd51c1a0b3200d667566ea53c4e05edb36c8d776b1db567e2d0a524"} Nov 25 10:54:13 crc kubenswrapper[4696]: I1125 10:54:13.921453 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"0d56e4965a783c6949a59b011b9faf78d61e63542b07316ff2ab67a4ebbfb5cb"} Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.239414 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364106 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsgjp\" (UniqueName: \"kubernetes.io/projected/7d786afc-ebdc-4806-9daf-86aa8d2563dd-kube-api-access-gsgjp\") pod \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364209 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-log-ovn\") pod \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364251 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run-ovn\") pod \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364328 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-scripts\") pod \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364355 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-additional-scripts\") pod \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364385 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run\") pod \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\" (UID: \"7d786afc-ebdc-4806-9daf-86aa8d2563dd\") " Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364732 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run" (OuterVolumeSpecName: "var-run") pod "7d786afc-ebdc-4806-9daf-86aa8d2563dd" (UID: "7d786afc-ebdc-4806-9daf-86aa8d2563dd"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364794 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7d786afc-ebdc-4806-9daf-86aa8d2563dd" (UID: "7d786afc-ebdc-4806-9daf-86aa8d2563dd"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.364864 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7d786afc-ebdc-4806-9daf-86aa8d2563dd" (UID: "7d786afc-ebdc-4806-9daf-86aa8d2563dd"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.365373 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7d786afc-ebdc-4806-9daf-86aa8d2563dd" (UID: "7d786afc-ebdc-4806-9daf-86aa8d2563dd"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.365638 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-scripts" (OuterVolumeSpecName: "scripts") pod "7d786afc-ebdc-4806-9daf-86aa8d2563dd" (UID: "7d786afc-ebdc-4806-9daf-86aa8d2563dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.370643 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d786afc-ebdc-4806-9daf-86aa8d2563dd-kube-api-access-gsgjp" (OuterVolumeSpecName: "kube-api-access-gsgjp") pod "7d786afc-ebdc-4806-9daf-86aa8d2563dd" (UID: "7d786afc-ebdc-4806-9daf-86aa8d2563dd"). InnerVolumeSpecName "kube-api-access-gsgjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.465884 4696 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.465920 4696 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.465933 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.465942 4696 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7d786afc-ebdc-4806-9daf-86aa8d2563dd-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.465954 4696 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d786afc-ebdc-4806-9daf-86aa8d2563dd-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.465964 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsgjp\" (UniqueName: \"kubernetes.io/projected/7d786afc-ebdc-4806-9daf-86aa8d2563dd-kube-api-access-gsgjp\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.509564 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jz4h7" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.934697 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"9a6912311fe80ad1474e77c4f443ee32e1f05519511bf7f525a04f4f18ce288e"} Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.937248 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jz4h7-config-z2gpm" event={"ID":"7d786afc-ebdc-4806-9daf-86aa8d2563dd","Type":"ContainerDied","Data":"1e347649ce5f0f16b39a637986ff5c3e6dbe356d7856e8220e6f3e86f99227e0"} Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.937274 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e347649ce5f0f16b39a637986ff5c3e6dbe356d7856e8220e6f3e86f99227e0" Nov 25 10:54:14 crc kubenswrapper[4696]: I1125 10:54:14.937296 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jz4h7-config-z2gpm" Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.359055 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jz4h7-config-z2gpm"] Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.372462 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jz4h7-config-z2gpm"] Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.827831 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.953082 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"33fdaeb60819911b1d1c212b79a06c86effb3b3926c9de10455237528bc8520d"} Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.953124 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"46d11868c600e46288adbcb5708d3efb42c603f8701a4e87c1928af6adf670a8"} Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.953134 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"888cb70d22fc01c65cb239fa203b20250f68ecc3d929fb201d953df6cbf747b5"} Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.953141 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"f788bbdc1feba1737bfdf1e3bc4b6988d0986eee9a5d502635b16b3d351e31b0"} Nov 25 10:54:15 crc kubenswrapper[4696]: I1125 10:54:15.988853 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.052700 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d786afc-ebdc-4806-9daf-86aa8d2563dd" path="/var/lib/kubelet/pods/7d786afc-ebdc-4806-9daf-86aa8d2563dd/volumes" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.421243 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qdf6p"] Nov 25 10:54:16 crc kubenswrapper[4696]: E1125 10:54:16.421596 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d786afc-ebdc-4806-9daf-86aa8d2563dd" containerName="ovn-config" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.421611 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d786afc-ebdc-4806-9daf-86aa8d2563dd" containerName="ovn-config" Nov 25 10:54:16 crc kubenswrapper[4696]: E1125 10:54:16.421627 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbeda82d-3956-47f7-95a8-6ffc25961e5d" containerName="swift-ring-rebalance" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.421635 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbeda82d-3956-47f7-95a8-6ffc25961e5d" containerName="swift-ring-rebalance" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.421817 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbeda82d-3956-47f7-95a8-6ffc25961e5d" containerName="swift-ring-rebalance" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.421839 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d786afc-ebdc-4806-9daf-86aa8d2563dd" containerName="ovn-config" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.422333 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.427710 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b1f9-account-create-4pwnx"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.428638 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.430184 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.443145 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b1f9-account-create-4pwnx"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.450570 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qdf6p"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.502603 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmr5x\" (UniqueName: \"kubernetes.io/projected/1c214c5b-9ca0-4499-814e-65a5fa45e42a-kube-api-access-fmr5x\") pod \"barbican-b1f9-account-create-4pwnx\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.502956 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c214c5b-9ca0-4499-814e-65a5fa45e42a-operator-scripts\") pod \"barbican-b1f9-account-create-4pwnx\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.502998 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzg42\" (UniqueName: \"kubernetes.io/projected/b5483447-69b5-485d-8ea2-2de150da78bb-kube-api-access-jzg42\") pod \"cinder-db-create-qdf6p\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.503067 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5483447-69b5-485d-8ea2-2de150da78bb-operator-scripts\") pod \"cinder-db-create-qdf6p\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.521475 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-bp7v2"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.522558 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.548509 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bp7v2"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.604676 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5483447-69b5-485d-8ea2-2de150da78bb-operator-scripts\") pod \"cinder-db-create-qdf6p\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.604726 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa48aa3d-d214-4690-8a29-a140a2c11d77-operator-scripts\") pod \"barbican-db-create-bp7v2\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.604804 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6p2s\" (UniqueName: \"kubernetes.io/projected/fa48aa3d-d214-4690-8a29-a140a2c11d77-kube-api-access-b6p2s\") pod \"barbican-db-create-bp7v2\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.604822 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmr5x\" (UniqueName: \"kubernetes.io/projected/1c214c5b-9ca0-4499-814e-65a5fa45e42a-kube-api-access-fmr5x\") pod \"barbican-b1f9-account-create-4pwnx\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.604854 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c214c5b-9ca0-4499-814e-65a5fa45e42a-operator-scripts\") pod \"barbican-b1f9-account-create-4pwnx\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.604879 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzg42\" (UniqueName: \"kubernetes.io/projected/b5483447-69b5-485d-8ea2-2de150da78bb-kube-api-access-jzg42\") pod \"cinder-db-create-qdf6p\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.605791 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5483447-69b5-485d-8ea2-2de150da78bb-operator-scripts\") pod \"cinder-db-create-qdf6p\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.606417 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c214c5b-9ca0-4499-814e-65a5fa45e42a-operator-scripts\") pod \"barbican-b1f9-account-create-4pwnx\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.625182 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-401d-account-create-ptvsr"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.626120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.628402 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.638237 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzg42\" (UniqueName: \"kubernetes.io/projected/b5483447-69b5-485d-8ea2-2de150da78bb-kube-api-access-jzg42\") pod \"cinder-db-create-qdf6p\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.669938 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-401d-account-create-ptvsr"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.697218 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmr5x\" (UniqueName: \"kubernetes.io/projected/1c214c5b-9ca0-4499-814e-65a5fa45e42a-kube-api-access-fmr5x\") pod \"barbican-b1f9-account-create-4pwnx\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.708603 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa48aa3d-d214-4690-8a29-a140a2c11d77-operator-scripts\") pod \"barbican-db-create-bp7v2\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.708706 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e389cdae-22d0-4285-8f08-d26bda099503-operator-scripts\") pod \"cinder-401d-account-create-ptvsr\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.708780 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6p2s\" (UniqueName: \"kubernetes.io/projected/fa48aa3d-d214-4690-8a29-a140a2c11d77-kube-api-access-b6p2s\") pod \"barbican-db-create-bp7v2\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.708840 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rctlt\" (UniqueName: \"kubernetes.io/projected/e389cdae-22d0-4285-8f08-d26bda099503-kube-api-access-rctlt\") pod \"cinder-401d-account-create-ptvsr\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.710069 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa48aa3d-d214-4690-8a29-a140a2c11d77-operator-scripts\") pod \"barbican-db-create-bp7v2\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.737933 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xtjbt"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.751597 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.752459 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6p2s\" (UniqueName: \"kubernetes.io/projected/fa48aa3d-d214-4690-8a29-a140a2c11d77-kube-api-access-b6p2s\") pod \"barbican-db-create-bp7v2\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.753536 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.765540 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.779822 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4xkzg"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.806625 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.816033 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.816354 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.816583 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.816745 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7vqt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.817732 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rctlt\" (UniqueName: \"kubernetes.io/projected/e389cdae-22d0-4285-8f08-d26bda099503-kube-api-access-rctlt\") pod \"cinder-401d-account-create-ptvsr\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.817802 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4phf\" (UniqueName: \"kubernetes.io/projected/68e586e5-acb4-4bc7-be27-545032778925-kube-api-access-r4phf\") pod \"neutron-db-create-xtjbt\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.817879 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e389cdae-22d0-4285-8f08-d26bda099503-operator-scripts\") pod \"cinder-401d-account-create-ptvsr\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.817916 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68e586e5-acb4-4bc7-be27-545032778925-operator-scripts\") pod \"neutron-db-create-xtjbt\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.831499 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e389cdae-22d0-4285-8f08-d26bda099503-operator-scripts\") pod \"cinder-401d-account-create-ptvsr\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.832080 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xtjbt"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.846843 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.850368 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4xkzg"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.896174 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d0d4-account-create-qhm9b"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.896341 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rctlt\" (UniqueName: \"kubernetes.io/projected/e389cdae-22d0-4285-8f08-d26bda099503-kube-api-access-rctlt\") pod \"cinder-401d-account-create-ptvsr\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.897275 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.902206 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.911513 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d0d4-account-create-qhm9b"] Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921436 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9hbj\" (UniqueName: \"kubernetes.io/projected/61fbc9f2-4e89-4d11-9d92-434700854f9f-kube-api-access-r9hbj\") pod \"neutron-d0d4-account-create-qhm9b\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921525 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68e586e5-acb4-4bc7-be27-545032778925-operator-scripts\") pod \"neutron-db-create-xtjbt\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921556 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-combined-ca-bundle\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921651 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-config-data\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921732 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4phf\" (UniqueName: \"kubernetes.io/projected/68e586e5-acb4-4bc7-be27-545032778925-kube-api-access-r4phf\") pod \"neutron-db-create-xtjbt\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921752 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cnjw\" (UniqueName: \"kubernetes.io/projected/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-kube-api-access-6cnjw\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.921772 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61fbc9f2-4e89-4d11-9d92-434700854f9f-operator-scripts\") pod \"neutron-d0d4-account-create-qhm9b\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.922458 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68e586e5-acb4-4bc7-be27-545032778925-operator-scripts\") pod \"neutron-db-create-xtjbt\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:16 crc kubenswrapper[4696]: I1125 10:54:16.939998 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4phf\" (UniqueName: \"kubernetes.io/projected/68e586e5-acb4-4bc7-be27-545032778925-kube-api-access-r4phf\") pod \"neutron-db-create-xtjbt\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.025510 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9hbj\" (UniqueName: \"kubernetes.io/projected/61fbc9f2-4e89-4d11-9d92-434700854f9f-kube-api-access-r9hbj\") pod \"neutron-d0d4-account-create-qhm9b\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.026284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-combined-ca-bundle\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.026375 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-config-data\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.026436 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cnjw\" (UniqueName: \"kubernetes.io/projected/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-kube-api-access-6cnjw\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.026464 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61fbc9f2-4e89-4d11-9d92-434700854f9f-operator-scripts\") pod \"neutron-d0d4-account-create-qhm9b\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.028389 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61fbc9f2-4e89-4d11-9d92-434700854f9f-operator-scripts\") pod \"neutron-d0d4-account-create-qhm9b\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.037151 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-config-data\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.044158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-combined-ca-bundle\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.054240 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cnjw\" (UniqueName: \"kubernetes.io/projected/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-kube-api-access-6cnjw\") pod \"keystone-db-sync-4xkzg\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.060447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9hbj\" (UniqueName: \"kubernetes.io/projected/61fbc9f2-4e89-4d11-9d92-434700854f9f-kube-api-access-r9hbj\") pod \"neutron-d0d4-account-create-qhm9b\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.082033 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.151872 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.164428 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:17 crc kubenswrapper[4696]: I1125 10:54:17.231770 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.310775 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d0d4-account-create-qhm9b"] Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.314571 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qdf6p"] Nov 25 10:54:18 crc kubenswrapper[4696]: W1125 10:54:18.353606 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61fbc9f2_4e89_4d11_9d92_434700854f9f.slice/crio-4f847e959ebf2a549446a9fcd2fbefe6ad820b2bef592a2b00b283f362521736 WatchSource:0}: Error finding container 4f847e959ebf2a549446a9fcd2fbefe6ad820b2bef592a2b00b283f362521736: Status 404 returned error can't find the container with id 4f847e959ebf2a549446a9fcd2fbefe6ad820b2bef592a2b00b283f362521736 Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.353845 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xtjbt"] Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.397897 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b1f9-account-create-4pwnx"] Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.414369 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4xkzg"] Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.435319 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-401d-account-create-ptvsr"] Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.456533 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bp7v2"] Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.979287 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-401d-account-create-ptvsr" event={"ID":"e389cdae-22d0-4285-8f08-d26bda099503","Type":"ContainerStarted","Data":"f097a231c9aaf17b4ff7aea39b71bffa2fa20b5dc3e03609074c84b8baa4a8c6"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.984259 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"ac861e367c420ce95301147963d53c9ec2284d289c52f83c16bdda3e2f1868a9"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.984296 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"317c9ab92d6b4c763f3a592878b1f5ce315e1d346646036e781c1fda52155512"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.986125 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d0d4-account-create-qhm9b" event={"ID":"61fbc9f2-4e89-4d11-9d92-434700854f9f","Type":"ContainerStarted","Data":"4f847e959ebf2a549446a9fcd2fbefe6ad820b2bef592a2b00b283f362521736"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.989861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bp7v2" event={"ID":"fa48aa3d-d214-4690-8a29-a140a2c11d77","Type":"ContainerStarted","Data":"872da4e1a3747968f825bac601440265b01f554123f9fd9fe0b75d058a1e1ddb"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.992462 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b1f9-account-create-4pwnx" event={"ID":"1c214c5b-9ca0-4499-814e-65a5fa45e42a","Type":"ContainerStarted","Data":"e8332f98c4246923175ad3a8fc28967629dbe8d7145c7c23d77cb8507acca34b"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.992519 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b1f9-account-create-4pwnx" event={"ID":"1c214c5b-9ca0-4499-814e-65a5fa45e42a","Type":"ContainerStarted","Data":"3cbb1ed26d557bfcc492d848493c50bae44f004ce91a6d24c105b334524aba32"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.994413 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qdf6p" event={"ID":"b5483447-69b5-485d-8ea2-2de150da78bb","Type":"ContainerStarted","Data":"1f446d8e3bd63349999c9b05cdcbc15544c2da06e464b752e13b25ef43e49bf8"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.994444 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qdf6p" event={"ID":"b5483447-69b5-485d-8ea2-2de150da78bb","Type":"ContainerStarted","Data":"2120e89743047cd141fa85d01506cd8decfd74651612b89ad65f0ca76b490453"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.995229 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4xkzg" event={"ID":"143d89a1-ef2c-4e93-8bd7-40e8a6044bca","Type":"ContainerStarted","Data":"4e13d9e79ae4ded95396ec02e103af1858807f766c71dfb6624ba6199ee3efef"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.996297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xtjbt" event={"ID":"68e586e5-acb4-4bc7-be27-545032778925","Type":"ContainerStarted","Data":"4c6005b993b3e6be023724f76800a565e233a0d13ad0c770ac718e2d07143625"} Nov 25 10:54:18 crc kubenswrapper[4696]: I1125 10:54:18.996338 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xtjbt" event={"ID":"68e586e5-acb4-4bc7-be27-545032778925","Type":"ContainerStarted","Data":"a923a905e7eb28433d9078185b1f3b269f7749740e5e247c857a027bf3f789e5"} Nov 25 10:54:19 crc kubenswrapper[4696]: I1125 10:54:19.023297 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-b1f9-account-create-4pwnx" podStartSLOduration=3.023278241 podStartE2EDuration="3.023278241s" podCreationTimestamp="2025-11-25 10:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:19.016705769 +0000 UTC m=+1055.830322366" watchObservedRunningTime="2025-11-25 10:54:19.023278241 +0000 UTC m=+1055.836894818" Nov 25 10:54:19 crc kubenswrapper[4696]: I1125 10:54:19.033908 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-xtjbt" podStartSLOduration=3.033888677 podStartE2EDuration="3.033888677s" podCreationTimestamp="2025-11-25 10:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:19.032421983 +0000 UTC m=+1055.846038590" watchObservedRunningTime="2025-11-25 10:54:19.033888677 +0000 UTC m=+1055.847505264" Nov 25 10:54:19 crc kubenswrapper[4696]: I1125 10:54:19.056161 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-qdf6p" podStartSLOduration=3.056142551 podStartE2EDuration="3.056142551s" podCreationTimestamp="2025-11-25 10:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:19.046559426 +0000 UTC m=+1055.860176013" watchObservedRunningTime="2025-11-25 10:54:19.056142551 +0000 UTC m=+1055.869759138" Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.017038 4696 generic.go:334] "Generic (PLEG): container finished" podID="b5483447-69b5-485d-8ea2-2de150da78bb" containerID="1f446d8e3bd63349999c9b05cdcbc15544c2da06e464b752e13b25ef43e49bf8" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.019091 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qdf6p" event={"ID":"b5483447-69b5-485d-8ea2-2de150da78bb","Type":"ContainerDied","Data":"1f446d8e3bd63349999c9b05cdcbc15544c2da06e464b752e13b25ef43e49bf8"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.046850 4696 generic.go:334] "Generic (PLEG): container finished" podID="e389cdae-22d0-4285-8f08-d26bda099503" containerID="1cdef9c1ffa9760a5e1318ecc1d79b3504841ca8e2ca6f7382c6f5fedf2011e5" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.055957 4696 generic.go:334] "Generic (PLEG): container finished" podID="68e586e5-acb4-4bc7-be27-545032778925" containerID="4c6005b993b3e6be023724f76800a565e233a0d13ad0c770ac718e2d07143625" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.064465 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-401d-account-create-ptvsr" event={"ID":"e389cdae-22d0-4285-8f08-d26bda099503","Type":"ContainerDied","Data":"1cdef9c1ffa9760a5e1318ecc1d79b3504841ca8e2ca6f7382c6f5fedf2011e5"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.064514 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xtjbt" event={"ID":"68e586e5-acb4-4bc7-be27-545032778925","Type":"ContainerDied","Data":"4c6005b993b3e6be023724f76800a565e233a0d13ad0c770ac718e2d07143625"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.067639 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"88589e635dc9bdd1c9b197b012f9a21b8128e78dca1d650dd03ccdfa72a3c003"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.067693 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"b64dd3395ee073d7fee587ab16d85ddb472d89946f4d9013e6af08b6de2d076d"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.067703 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"e7c23efb7224baeab6abe48392780e3a9c0e77aa1d94d7c1bd0c929ed4eb7002"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.067711 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"584b0ee72638a378eea58eea361be595b11bcc92e79e7553dcfac03e6455e2fc"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.072863 4696 generic.go:334] "Generic (PLEG): container finished" podID="61fbc9f2-4e89-4d11-9d92-434700854f9f" containerID="66a23011e15065483610e5debb6babcdae1dfb3c4d1bc930ca2d74b7dd80fa29" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.072932 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d0d4-account-create-qhm9b" event={"ID":"61fbc9f2-4e89-4d11-9d92-434700854f9f","Type":"ContainerDied","Data":"66a23011e15065483610e5debb6babcdae1dfb3c4d1bc930ca2d74b7dd80fa29"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.077589 4696 generic.go:334] "Generic (PLEG): container finished" podID="fa48aa3d-d214-4690-8a29-a140a2c11d77" containerID="0a3edf2ba429339228ed5782c0e0039a92d3dfaf39806cc4e9f1515032854441" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.077648 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bp7v2" event={"ID":"fa48aa3d-d214-4690-8a29-a140a2c11d77","Type":"ContainerDied","Data":"0a3edf2ba429339228ed5782c0e0039a92d3dfaf39806cc4e9f1515032854441"} Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.082727 4696 generic.go:334] "Generic (PLEG): container finished" podID="1c214c5b-9ca0-4499-814e-65a5fa45e42a" containerID="e8332f98c4246923175ad3a8fc28967629dbe8d7145c7c23d77cb8507acca34b" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4696]: I1125 10:54:20.082772 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b1f9-account-create-4pwnx" event={"ID":"1c214c5b-9ca0-4499-814e-65a5fa45e42a","Type":"ContainerDied","Data":"e8332f98c4246923175ad3a8fc28967629dbe8d7145c7c23d77cb8507acca34b"} Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.097359 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e87af5be-8499-4303-a158-8c9d50c2db73","Type":"ContainerStarted","Data":"eada619cd4af2a4b5b65a20a171f53f04c4ef951b8f682c0203b02491e547fcf"} Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.183179 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=25.365441294 podStartE2EDuration="31.183159296s" podCreationTimestamp="2025-11-25 10:53:50 +0000 UTC" firstStartedPulling="2025-11-25 10:54:12.095258163 +0000 UTC m=+1048.908874750" lastFinishedPulling="2025-11-25 10:54:17.912976155 +0000 UTC m=+1054.726592752" observedRunningTime="2025-11-25 10:54:21.183068674 +0000 UTC m=+1057.996685261" watchObservedRunningTime="2025-11-25 10:54:21.183159296 +0000 UTC m=+1057.996775883" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.475234 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-wtvbr"] Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.482463 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.488412 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.518964 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-wtvbr"] Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.567904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.567958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-config\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.567986 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.568023 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhk6h\" (UniqueName: \"kubernetes.io/projected/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-kube-api-access-zhk6h\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.568045 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.568062 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.671055 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhk6h\" (UniqueName: \"kubernetes.io/projected/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-kube-api-access-zhk6h\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.671470 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.671487 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.671576 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.671604 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-config\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.671630 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.673501 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.674207 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.674297 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.674879 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-config\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.675533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.704815 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhk6h\" (UniqueName: \"kubernetes.io/projected/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-kube-api-access-zhk6h\") pod \"dnsmasq-dns-5c79d794d7-wtvbr\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:21 crc kubenswrapper[4696]: I1125 10:54:21.811628 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.142278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d0d4-account-create-qhm9b" event={"ID":"61fbc9f2-4e89-4d11-9d92-434700854f9f","Type":"ContainerDied","Data":"4f847e959ebf2a549446a9fcd2fbefe6ad820b2bef592a2b00b283f362521736"} Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.142687 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f847e959ebf2a549446a9fcd2fbefe6ad820b2bef592a2b00b283f362521736" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.145060 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bp7v2" event={"ID":"fa48aa3d-d214-4690-8a29-a140a2c11d77","Type":"ContainerDied","Data":"872da4e1a3747968f825bac601440265b01f554123f9fd9fe0b75d058a1e1ddb"} Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.145087 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="872da4e1a3747968f825bac601440265b01f554123f9fd9fe0b75d058a1e1ddb" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.151473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b1f9-account-create-4pwnx" event={"ID":"1c214c5b-9ca0-4499-814e-65a5fa45e42a","Type":"ContainerDied","Data":"3cbb1ed26d557bfcc492d848493c50bae44f004ce91a6d24c105b334524aba32"} Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.151517 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cbb1ed26d557bfcc492d848493c50bae44f004ce91a6d24c105b334524aba32" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.153002 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qdf6p" event={"ID":"b5483447-69b5-485d-8ea2-2de150da78bb","Type":"ContainerDied","Data":"2120e89743047cd141fa85d01506cd8decfd74651612b89ad65f0ca76b490453"} Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.153049 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2120e89743047cd141fa85d01506cd8decfd74651612b89ad65f0ca76b490453" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.155657 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xtjbt" event={"ID":"68e586e5-acb4-4bc7-be27-545032778925","Type":"ContainerDied","Data":"a923a905e7eb28433d9078185b1f3b269f7749740e5e247c857a027bf3f789e5"} Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.155706 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a923a905e7eb28433d9078185b1f3b269f7749740e5e247c857a027bf3f789e5" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.157933 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-401d-account-create-ptvsr" event={"ID":"e389cdae-22d0-4285-8f08-d26bda099503","Type":"ContainerDied","Data":"f097a231c9aaf17b4ff7aea39b71bffa2fa20b5dc3e03609074c84b8baa4a8c6"} Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.157962 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f097a231c9aaf17b4ff7aea39b71bffa2fa20b5dc3e03609074c84b8baa4a8c6" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.190842 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.231376 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4phf\" (UniqueName: \"kubernetes.io/projected/68e586e5-acb4-4bc7-be27-545032778925-kube-api-access-r4phf\") pod \"68e586e5-acb4-4bc7-be27-545032778925\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.231484 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68e586e5-acb4-4bc7-be27-545032778925-operator-scripts\") pod \"68e586e5-acb4-4bc7-be27-545032778925\" (UID: \"68e586e5-acb4-4bc7-be27-545032778925\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.233272 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68e586e5-acb4-4bc7-be27-545032778925-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68e586e5-acb4-4bc7-be27-545032778925" (UID: "68e586e5-acb4-4bc7-be27-545032778925"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.249203 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68e586e5-acb4-4bc7-be27-545032778925-kube-api-access-r4phf" (OuterVolumeSpecName: "kube-api-access-r4phf") pod "68e586e5-acb4-4bc7-be27-545032778925" (UID: "68e586e5-acb4-4bc7-be27-545032778925"). InnerVolumeSpecName "kube-api-access-r4phf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.317018 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.331965 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.333181 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61fbc9f2-4e89-4d11-9d92-434700854f9f-operator-scripts\") pod \"61fbc9f2-4e89-4d11-9d92-434700854f9f\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.333224 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9hbj\" (UniqueName: \"kubernetes.io/projected/61fbc9f2-4e89-4d11-9d92-434700854f9f-kube-api-access-r9hbj\") pod \"61fbc9f2-4e89-4d11-9d92-434700854f9f\" (UID: \"61fbc9f2-4e89-4d11-9d92-434700854f9f\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.333552 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4phf\" (UniqueName: \"kubernetes.io/projected/68e586e5-acb4-4bc7-be27-545032778925-kube-api-access-r4phf\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.333565 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68e586e5-acb4-4bc7-be27-545032778925-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.334150 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fbc9f2-4e89-4d11-9d92-434700854f9f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61fbc9f2-4e89-4d11-9d92-434700854f9f" (UID: "61fbc9f2-4e89-4d11-9d92-434700854f9f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.342541 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61fbc9f2-4e89-4d11-9d92-434700854f9f-kube-api-access-r9hbj" (OuterVolumeSpecName: "kube-api-access-r9hbj") pod "61fbc9f2-4e89-4d11-9d92-434700854f9f" (UID: "61fbc9f2-4e89-4d11-9d92-434700854f9f"). InnerVolumeSpecName "kube-api-access-r9hbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.346586 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.359786 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.383171 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434249 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rctlt\" (UniqueName: \"kubernetes.io/projected/e389cdae-22d0-4285-8f08-d26bda099503-kube-api-access-rctlt\") pod \"e389cdae-22d0-4285-8f08-d26bda099503\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434578 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6p2s\" (UniqueName: \"kubernetes.io/projected/fa48aa3d-d214-4690-8a29-a140a2c11d77-kube-api-access-b6p2s\") pod \"fa48aa3d-d214-4690-8a29-a140a2c11d77\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434604 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzg42\" (UniqueName: \"kubernetes.io/projected/b5483447-69b5-485d-8ea2-2de150da78bb-kube-api-access-jzg42\") pod \"b5483447-69b5-485d-8ea2-2de150da78bb\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434633 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5483447-69b5-485d-8ea2-2de150da78bb-operator-scripts\") pod \"b5483447-69b5-485d-8ea2-2de150da78bb\" (UID: \"b5483447-69b5-485d-8ea2-2de150da78bb\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434833 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c214c5b-9ca0-4499-814e-65a5fa45e42a-operator-scripts\") pod \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434864 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa48aa3d-d214-4690-8a29-a140a2c11d77-operator-scripts\") pod \"fa48aa3d-d214-4690-8a29-a140a2c11d77\" (UID: \"fa48aa3d-d214-4690-8a29-a140a2c11d77\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434901 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e389cdae-22d0-4285-8f08-d26bda099503-operator-scripts\") pod \"e389cdae-22d0-4285-8f08-d26bda099503\" (UID: \"e389cdae-22d0-4285-8f08-d26bda099503\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.434930 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmr5x\" (UniqueName: \"kubernetes.io/projected/1c214c5b-9ca0-4499-814e-65a5fa45e42a-kube-api-access-fmr5x\") pod \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\" (UID: \"1c214c5b-9ca0-4499-814e-65a5fa45e42a\") " Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.435233 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61fbc9f2-4e89-4d11-9d92-434700854f9f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.435256 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9hbj\" (UniqueName: \"kubernetes.io/projected/61fbc9f2-4e89-4d11-9d92-434700854f9f-kube-api-access-r9hbj\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.436591 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5483447-69b5-485d-8ea2-2de150da78bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5483447-69b5-485d-8ea2-2de150da78bb" (UID: "b5483447-69b5-485d-8ea2-2de150da78bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.436623 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e389cdae-22d0-4285-8f08-d26bda099503-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e389cdae-22d0-4285-8f08-d26bda099503" (UID: "e389cdae-22d0-4285-8f08-d26bda099503"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.436649 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa48aa3d-d214-4690-8a29-a140a2c11d77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa48aa3d-d214-4690-8a29-a140a2c11d77" (UID: "fa48aa3d-d214-4690-8a29-a140a2c11d77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.437525 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c214c5b-9ca0-4499-814e-65a5fa45e42a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c214c5b-9ca0-4499-814e-65a5fa45e42a" (UID: "1c214c5b-9ca0-4499-814e-65a5fa45e42a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.438059 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e389cdae-22d0-4285-8f08-d26bda099503-kube-api-access-rctlt" (OuterVolumeSpecName: "kube-api-access-rctlt") pod "e389cdae-22d0-4285-8f08-d26bda099503" (UID: "e389cdae-22d0-4285-8f08-d26bda099503"). InnerVolumeSpecName "kube-api-access-rctlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.438692 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa48aa3d-d214-4690-8a29-a140a2c11d77-kube-api-access-b6p2s" (OuterVolumeSpecName: "kube-api-access-b6p2s") pod "fa48aa3d-d214-4690-8a29-a140a2c11d77" (UID: "fa48aa3d-d214-4690-8a29-a140a2c11d77"). InnerVolumeSpecName "kube-api-access-b6p2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.439136 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5483447-69b5-485d-8ea2-2de150da78bb-kube-api-access-jzg42" (OuterVolumeSpecName: "kube-api-access-jzg42") pod "b5483447-69b5-485d-8ea2-2de150da78bb" (UID: "b5483447-69b5-485d-8ea2-2de150da78bb"). InnerVolumeSpecName "kube-api-access-jzg42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.439181 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c214c5b-9ca0-4499-814e-65a5fa45e42a-kube-api-access-fmr5x" (OuterVolumeSpecName: "kube-api-access-fmr5x") pod "1c214c5b-9ca0-4499-814e-65a5fa45e42a" (UID: "1c214c5b-9ca0-4499-814e-65a5fa45e42a"). InnerVolumeSpecName "kube-api-access-fmr5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.508173 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-wtvbr"] Nov 25 10:54:25 crc kubenswrapper[4696]: W1125 10:54:25.511229 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6aac8f88_5b55_4f25_8cee_6b20fa9faecd.slice/crio-76d5c07cd580b6165ef07a66408ed480c7835e40e5b1d22e3c23fafc645978ee WatchSource:0}: Error finding container 76d5c07cd580b6165ef07a66408ed480c7835e40e5b1d22e3c23fafc645978ee: Status 404 returned error can't find the container with id 76d5c07cd580b6165ef07a66408ed480c7835e40e5b1d22e3c23fafc645978ee Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.536824 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rctlt\" (UniqueName: \"kubernetes.io/projected/e389cdae-22d0-4285-8f08-d26bda099503-kube-api-access-rctlt\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.536905 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6p2s\" (UniqueName: \"kubernetes.io/projected/fa48aa3d-d214-4690-8a29-a140a2c11d77-kube-api-access-b6p2s\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.537027 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzg42\" (UniqueName: \"kubernetes.io/projected/b5483447-69b5-485d-8ea2-2de150da78bb-kube-api-access-jzg42\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.537383 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5483447-69b5-485d-8ea2-2de150da78bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.537499 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c214c5b-9ca0-4499-814e-65a5fa45e42a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.537621 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa48aa3d-d214-4690-8a29-a140a2c11d77-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.537734 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e389cdae-22d0-4285-8f08-d26bda099503-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4696]: I1125 10:54:25.537814 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmr5x\" (UniqueName: \"kubernetes.io/projected/1c214c5b-9ca0-4499-814e-65a5fa45e42a-kube-api-access-fmr5x\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.167584 4696 generic.go:334] "Generic (PLEG): container finished" podID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerID="0f4b495f33536cabfe56fbf85bd87cd45f4b9603bbffaeaca651bbdcbeb8d9e8" exitCode=0 Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.167635 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" event={"ID":"6aac8f88-5b55-4f25-8cee-6b20fa9faecd","Type":"ContainerDied","Data":"0f4b495f33536cabfe56fbf85bd87cd45f4b9603bbffaeaca651bbdcbeb8d9e8"} Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.167992 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" event={"ID":"6aac8f88-5b55-4f25-8cee-6b20fa9faecd","Type":"ContainerStarted","Data":"76d5c07cd580b6165ef07a66408ed480c7835e40e5b1d22e3c23fafc645978ee"} Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.170294 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ljbtx" event={"ID":"bc2593ea-7985-4050-8000-2be38f4b488c","Type":"ContainerStarted","Data":"62be1a71b1f817ac50a79718ab71a3fa8ac78f3bd0933da18172e94b2d922df4"} Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.172181 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xtjbt" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.173091 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4xkzg" event={"ID":"143d89a1-ef2c-4e93-8bd7-40e8a6044bca","Type":"ContainerStarted","Data":"d61549e48123092e2468424e4ab44901eb7c82dda018ed09824954bc88803735"} Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.173164 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d0d4-account-create-qhm9b" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.173526 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-401d-account-create-ptvsr" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.173903 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bp7v2" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.174306 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b1f9-account-create-4pwnx" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.174731 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qdf6p" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.252534 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ljbtx" podStartSLOduration=1.840406798 podStartE2EDuration="32.252517068s" podCreationTimestamp="2025-11-25 10:53:54 +0000 UTC" firstStartedPulling="2025-11-25 10:53:55.080044607 +0000 UTC m=+1031.893661194" lastFinishedPulling="2025-11-25 10:54:25.492154877 +0000 UTC m=+1062.305771464" observedRunningTime="2025-11-25 10:54:26.204717449 +0000 UTC m=+1063.018334046" watchObservedRunningTime="2025-11-25 10:54:26.252517068 +0000 UTC m=+1063.066133655" Nov 25 10:54:26 crc kubenswrapper[4696]: I1125 10:54:26.265330 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4xkzg" podStartSLOduration=3.728767485 podStartE2EDuration="10.265308751s" podCreationTimestamp="2025-11-25 10:54:16 +0000 UTC" firstStartedPulling="2025-11-25 10:54:18.514574166 +0000 UTC m=+1055.328190753" lastFinishedPulling="2025-11-25 10:54:25.051115432 +0000 UTC m=+1061.864732019" observedRunningTime="2025-11-25 10:54:26.240947152 +0000 UTC m=+1063.054563749" watchObservedRunningTime="2025-11-25 10:54:26.265308751 +0000 UTC m=+1063.078925348" Nov 25 10:54:27 crc kubenswrapper[4696]: I1125 10:54:27.192130 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" event={"ID":"6aac8f88-5b55-4f25-8cee-6b20fa9faecd","Type":"ContainerStarted","Data":"8f50f81db271f4f47fc5b1a6901dce23cc6399e8c12abe5901b52750df59aa40"} Nov 25 10:54:27 crc kubenswrapper[4696]: I1125 10:54:27.192355 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:27 crc kubenswrapper[4696]: I1125 10:54:27.217415 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podStartSLOduration=6.217400584 podStartE2EDuration="6.217400584s" podCreationTimestamp="2025-11-25 10:54:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:27.214772903 +0000 UTC m=+1064.028389490" watchObservedRunningTime="2025-11-25 10:54:27.217400584 +0000 UTC m=+1064.031017171" Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.218347 4696 generic.go:334] "Generic (PLEG): container finished" podID="143d89a1-ef2c-4e93-8bd7-40e8a6044bca" containerID="d61549e48123092e2468424e4ab44901eb7c82dda018ed09824954bc88803735" exitCode=0 Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.218457 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4xkzg" event={"ID":"143d89a1-ef2c-4e93-8bd7-40e8a6044bca","Type":"ContainerDied","Data":"d61549e48123092e2468424e4ab44901eb7c82dda018ed09824954bc88803735"} Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.802474 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.802994 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.803077 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.804284 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33cc4b473fa3af53c05eee091add2db7db56314c26dffae135812139c5ec3acf"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:54:30 crc kubenswrapper[4696]: I1125 10:54:30.804410 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://33cc4b473fa3af53c05eee091add2db7db56314c26dffae135812139c5ec3acf" gracePeriod=600 Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.228817 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="33cc4b473fa3af53c05eee091add2db7db56314c26dffae135812139c5ec3acf" exitCode=0 Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.228872 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"33cc4b473fa3af53c05eee091add2db7db56314c26dffae135812139c5ec3acf"} Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.229263 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769"} Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.229324 4696 scope.go:117] "RemoveContainer" containerID="e7253a2ef071b30ac37f2b8fc9846ceb3e4a68877fb90b28472f1abfa37caf36" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.535997 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.537477 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cnjw\" (UniqueName: \"kubernetes.io/projected/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-kube-api-access-6cnjw\") pod \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.537572 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-combined-ca-bundle\") pod \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.537631 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-config-data\") pod \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\" (UID: \"143d89a1-ef2c-4e93-8bd7-40e8a6044bca\") " Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.546964 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-kube-api-access-6cnjw" (OuterVolumeSpecName: "kube-api-access-6cnjw") pod "143d89a1-ef2c-4e93-8bd7-40e8a6044bca" (UID: "143d89a1-ef2c-4e93-8bd7-40e8a6044bca"). InnerVolumeSpecName "kube-api-access-6cnjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.578349 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "143d89a1-ef2c-4e93-8bd7-40e8a6044bca" (UID: "143d89a1-ef2c-4e93-8bd7-40e8a6044bca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.600547 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-config-data" (OuterVolumeSpecName: "config-data") pod "143d89a1-ef2c-4e93-8bd7-40e8a6044bca" (UID: "143d89a1-ef2c-4e93-8bd7-40e8a6044bca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.639327 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.639585 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.639597 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cnjw\" (UniqueName: \"kubernetes.io/projected/143d89a1-ef2c-4e93-8bd7-40e8a6044bca-kube-api-access-6cnjw\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.813737 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.869340 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-ljbcg"] Nov 25 10:54:31 crc kubenswrapper[4696]: I1125 10:54:31.869583 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerName="dnsmasq-dns" containerID="cri-o://a196b7680557afe9147c63a12d5a58855670014218b4c017d107fdab61427c25" gracePeriod=10 Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.241723 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4xkzg" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.241729 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4xkzg" event={"ID":"143d89a1-ef2c-4e93-8bd7-40e8a6044bca","Type":"ContainerDied","Data":"4e13d9e79ae4ded95396ec02e103af1858807f766c71dfb6624ba6199ee3efef"} Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.242150 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e13d9e79ae4ded95396ec02e103af1858807f766c71dfb6624ba6199ee3efef" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.252571 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" event={"ID":"cbc24b3a-9e3a-4653-8617-45eac69ac198","Type":"ContainerDied","Data":"a196b7680557afe9147c63a12d5a58855670014218b4c017d107fdab61427c25"} Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.252859 4696 generic.go:334] "Generic (PLEG): container finished" podID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerID="a196b7680557afe9147c63a12d5a58855670014218b4c017d107fdab61427c25" exitCode=0 Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.547753 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sbcmj"] Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548082 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c214c5b-9ca0-4499-814e-65a5fa45e42a" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548099 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c214c5b-9ca0-4499-814e-65a5fa45e42a" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548117 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143d89a1-ef2c-4e93-8bd7-40e8a6044bca" containerName="keystone-db-sync" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548125 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="143d89a1-ef2c-4e93-8bd7-40e8a6044bca" containerName="keystone-db-sync" Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548150 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa48aa3d-d214-4690-8a29-a140a2c11d77" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548161 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa48aa3d-d214-4690-8a29-a140a2c11d77" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548172 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5483447-69b5-485d-8ea2-2de150da78bb" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548178 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5483447-69b5-485d-8ea2-2de150da78bb" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548190 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e389cdae-22d0-4285-8f08-d26bda099503" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548195 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e389cdae-22d0-4285-8f08-d26bda099503" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548211 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68e586e5-acb4-4bc7-be27-545032778925" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548216 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="68e586e5-acb4-4bc7-be27-545032778925" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: E1125 10:54:32.548230 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fbc9f2-4e89-4d11-9d92-434700854f9f" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548235 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fbc9f2-4e89-4d11-9d92-434700854f9f" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548387 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa48aa3d-d214-4690-8a29-a140a2c11d77" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548408 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="68e586e5-acb4-4bc7-be27-545032778925" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548418 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5483447-69b5-485d-8ea2-2de150da78bb" containerName="mariadb-database-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548432 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fbc9f2-4e89-4d11-9d92-434700854f9f" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548446 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="143d89a1-ef2c-4e93-8bd7-40e8a6044bca" containerName="keystone-db-sync" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548454 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c214c5b-9ca0-4499-814e-65a5fa45e42a" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.548466 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e389cdae-22d0-4285-8f08-d26bda099503" containerName="mariadb-account-create" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.549020 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.553350 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.564517 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.569458 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.569709 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7vqt" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.571430 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.593205 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sbcmj"] Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.642808 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8kf4c"] Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.645999 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.672932 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-config-data\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.673527 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-scripts\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.673719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht79p\" (UniqueName: \"kubernetes.io/projected/7562e866-74fd-44f7-9d64-b38073fb6fa3-kube-api-access-ht79p\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.673811 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-fernet-keys\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.674056 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-credential-keys\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.674243 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-combined-ca-bundle\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.730758 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8kf4c"] Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775570 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-scripts\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775644 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht79p\" (UniqueName: \"kubernetes.io/projected/7562e866-74fd-44f7-9d64-b38073fb6fa3-kube-api-access-ht79p\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775684 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-fernet-keys\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775721 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-config\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775782 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-credential-keys\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775802 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775822 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq8vb\" (UniqueName: \"kubernetes.io/projected/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-kube-api-access-fq8vb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775846 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775868 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775887 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-combined-ca-bundle\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775928 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-config-data\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.775957 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-svc\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.810683 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-config-data\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.832290 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-combined-ca-bundle\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.836292 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-scripts\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.836565 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7557dc88f-7gn6r"] Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.838782 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-credential-keys\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.844232 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.845586 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht79p\" (UniqueName: \"kubernetes.io/projected/7562e866-74fd-44f7-9d64-b38073fb6fa3-kube-api-access-ht79p\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.857144 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.857381 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.857485 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-cnbgj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.857580 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.877593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.887977 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.888097 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq8vb\" (UniqueName: \"kubernetes.io/projected/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-kube-api-access-fq8vb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.888225 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.888288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.888513 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-svc\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.888753 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-config\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.889083 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.904021 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-config\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.917530 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.925755 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7557dc88f-7gn6r"] Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.928376 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-svc\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.929004 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-fernet-keys\") pod \"keystone-bootstrap-sbcmj\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.966433 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq8vb\" (UniqueName: \"kubernetes.io/projected/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-kube-api-access-fq8vb\") pod \"dnsmasq-dns-5b868669f-8kf4c\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.992837 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-config-data\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.993094 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-scripts\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.993230 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f437b119-d3e9-4145-9323-56fd522cd092-horizon-secret-key\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.993347 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbqtl\" (UniqueName: \"kubernetes.io/projected/f437b119-d3e9-4145-9323-56fd522cd092-kube-api-access-hbqtl\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:32 crc kubenswrapper[4696]: I1125 10:54:32.993424 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f437b119-d3e9-4145-9323-56fd522cd092-logs\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.052760 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.054813 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.073175 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.086390 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.096755 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-config-data\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.096839 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-scripts\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.096864 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f437b119-d3e9-4145-9323-56fd522cd092-horizon-secret-key\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.096930 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbqtl\" (UniqueName: \"kubernetes.io/projected/f437b119-d3e9-4145-9323-56fd522cd092-kube-api-access-hbqtl\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.096953 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f437b119-d3e9-4145-9323-56fd522cd092-logs\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.097358 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f437b119-d3e9-4145-9323-56fd522cd092-logs\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.098283 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-config-data\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.105785 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.106243 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-scripts\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.114264 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5vqmj"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.125444 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.123332 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f437b119-d3e9-4145-9323-56fd522cd092-horizon-secret-key\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.137518 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.139475 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-26hbb" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.169094 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7qtsq"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.170407 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.174006 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.174217 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.174321 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jl954" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.179105 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.203532 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-log-httpd\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.203576 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9bm4\" (UniqueName: \"kubernetes.io/projected/96044f36-992b-44ce-89b4-a059b1efc117-kube-api-access-b9bm4\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.203618 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.203648 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-run-httpd\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.203687 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-scripts\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.204118 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.205005 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-config-data\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.205032 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.220317 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbqtl\" (UniqueName: \"kubernetes.io/projected/f437b119-d3e9-4145-9323-56fd522cd092-kube-api-access-hbqtl\") pod \"horizon-7557dc88f-7gn6r\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.246737 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7qtsq"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.274641 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5vqmj"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.307506 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-config-data\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.307553 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.307579 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9353d93a-d886-43d0-a4d4-6ad17a28f52a-etc-machine-id\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.307607 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-db-sync-config-data\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.307624 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-scripts\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.307646 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5xg9\" (UniqueName: \"kubernetes.io/projected/cacd4fa8-b70a-451d-92a9-40e53ea0a406-kube-api-access-k5xg9\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.308902 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-log-httpd\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.308945 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9bm4\" (UniqueName: \"kubernetes.io/projected/96044f36-992b-44ce-89b4-a059b1efc117-kube-api-access-b9bm4\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.308997 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309038 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-run-httpd\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309056 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-combined-ca-bundle\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309096 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-combined-ca-bundle\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309121 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-scripts\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309191 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-config-data\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309240 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skp7d\" (UniqueName: \"kubernetes.io/projected/9353d93a-d886-43d0-a4d4-6ad17a28f52a-kube-api-access-skp7d\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.309274 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-db-sync-config-data\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.318869 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.319437 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-run-httpd\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.330528 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.335160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.335442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-scripts\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.349335 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-log-httpd\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.350599 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-config-data\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.385711 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nqj4p"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.391035 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.391908 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dq4dk"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.393283 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410446 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9353d93a-d886-43d0-a4d4-6ad17a28f52a-etc-machine-id\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410515 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-scripts\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410545 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-db-sync-config-data\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410576 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5xg9\" (UniqueName: \"kubernetes.io/projected/cacd4fa8-b70a-451d-92a9-40e53ea0a406-kube-api-access-k5xg9\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410643 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-combined-ca-bundle\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410690 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-combined-ca-bundle\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410735 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-config-data\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410770 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skp7d\" (UniqueName: \"kubernetes.io/projected/9353d93a-d886-43d0-a4d4-6ad17a28f52a-kube-api-access-skp7d\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.410801 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-db-sync-config-data\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.420802 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9353d93a-d886-43d0-a4d4-6ad17a28f52a-etc-machine-id\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.429221 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-combined-ca-bundle\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.429911 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-scripts\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.435221 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-db-sync-config-data\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.435968 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.436509 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-combined-ca-bundle\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.437101 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.438614 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-db-sync-config-data\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.445311 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-config-data\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.448004 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.448223 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-gmkwd" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.448353 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-445jl" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.448543 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.459039 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.471415 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9bm4\" (UniqueName: \"kubernetes.io/projected/96044f36-992b-44ce-89b4-a059b1efc117-kube-api-access-b9bm4\") pod \"ceilometer-0\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.474816 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5xg9\" (UniqueName: \"kubernetes.io/projected/cacd4fa8-b70a-451d-92a9-40e53ea0a406-kube-api-access-k5xg9\") pod \"barbican-db-sync-5vqmj\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.476024 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dq4dk"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.477055 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skp7d\" (UniqueName: \"kubernetes.io/projected/9353d93a-d886-43d0-a4d4-6ad17a28f52a-kube-api-access-skp7d\") pod \"cinder-db-sync-7qtsq\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.486099 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nqj4p"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.510648 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.520587 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8kf4c"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538028 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-logs\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538114 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2c2c\" (UniqueName: \"kubernetes.io/projected/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-kube-api-access-d2c2c\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538197 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-combined-ca-bundle\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538227 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-scripts\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538360 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-combined-ca-bundle\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538429 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-config-data\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538455 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r86gc\" (UniqueName: \"kubernetes.io/projected/3a052a79-e780-430c-a049-cc01418ba194-kube-api-access-r86gc\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.538488 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-config\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.583034 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d6957f485-4l99x"] Nov 25 10:54:33 crc kubenswrapper[4696]: E1125 10:54:33.583365 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerName="dnsmasq-dns" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.583376 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerName="dnsmasq-dns" Nov 25 10:54:33 crc kubenswrapper[4696]: E1125 10:54:33.583391 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerName="init" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.583397 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerName="init" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.583602 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" containerName="dnsmasq-dns" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.584365 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.611901 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d6957f485-4l99x"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.641484 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-config\") pod \"cbc24b3a-9e3a-4653-8617-45eac69ac198\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.641561 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-sb\") pod \"cbc24b3a-9e3a-4653-8617-45eac69ac198\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.641626 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-dns-svc\") pod \"cbc24b3a-9e3a-4653-8617-45eac69ac198\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.641647 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-nb\") pod \"cbc24b3a-9e3a-4653-8617-45eac69ac198\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.641710 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbvfq\" (UniqueName: \"kubernetes.io/projected/cbc24b3a-9e3a-4653-8617-45eac69ac198-kube-api-access-nbvfq\") pod \"cbc24b3a-9e3a-4653-8617-45eac69ac198\" (UID: \"cbc24b3a-9e3a-4653-8617-45eac69ac198\") " Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642019 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-combined-ca-bundle\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642081 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-config-data\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642102 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r86gc\" (UniqueName: \"kubernetes.io/projected/3a052a79-e780-430c-a049-cc01418ba194-kube-api-access-r86gc\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642124 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-config\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642176 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-logs\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642199 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2c2c\" (UniqueName: \"kubernetes.io/projected/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-kube-api-access-d2c2c\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642245 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-combined-ca-bundle\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.642263 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-scripts\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.659400 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-logs\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.659722 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.663241 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-scripts\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.664959 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-7d95t"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.666376 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.685788 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-config\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.687214 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.687291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-combined-ca-bundle\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.703393 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-7d95t"] Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.706533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-combined-ca-bundle\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.713581 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-config-data\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.729838 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc24b3a-9e3a-4653-8617-45eac69ac198-kube-api-access-nbvfq" (OuterVolumeSpecName: "kube-api-access-nbvfq") pod "cbc24b3a-9e3a-4653-8617-45eac69ac198" (UID: "cbc24b3a-9e3a-4653-8617-45eac69ac198"). InnerVolumeSpecName "kube-api-access-nbvfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.749570 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d38303a3-737e-44cb-a536-6f618e6fa272-horizon-secret-key\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.749637 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-config-data\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.749683 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s844k\" (UniqueName: \"kubernetes.io/projected/d38303a3-737e-44cb-a536-6f618e6fa272-kube-api-access-s844k\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.749739 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-scripts\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.749773 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38303a3-737e-44cb-a536-6f618e6fa272-logs\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.755556 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbvfq\" (UniqueName: \"kubernetes.io/projected/cbc24b3a-9e3a-4653-8617-45eac69ac198-kube-api-access-nbvfq\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.771583 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r86gc\" (UniqueName: \"kubernetes.io/projected/3a052a79-e780-430c-a049-cc01418ba194-kube-api-access-r86gc\") pod \"neutron-db-sync-nqj4p\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.775539 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2c2c\" (UniqueName: \"kubernetes.io/projected/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-kube-api-access-d2c2c\") pod \"placement-db-sync-dq4dk\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.876633 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cbc24b3a-9e3a-4653-8617-45eac69ac198" (UID: "cbc24b3a-9e3a-4653-8617-45eac69ac198"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.903225 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d38303a3-737e-44cb-a536-6f618e6fa272-horizon-secret-key\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.919288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-config-data\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.922027 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-config-data\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.929258 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d38303a3-737e-44cb-a536-6f618e6fa272-horizon-secret-key\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.929282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s844k\" (UniqueName: \"kubernetes.io/projected/d38303a3-737e-44cb-a536-6f618e6fa272-kube-api-access-s844k\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.930723 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-config\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.930820 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.930873 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.930955 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-scripts\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.931049 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-svc\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.931107 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zp6m\" (UniqueName: \"kubernetes.io/projected/281499bc-198d-46c6-82e6-ee9260834650-kube-api-access-9zp6m\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.931136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38303a3-737e-44cb-a536-6f618e6fa272-logs\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.931234 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.931354 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.931636 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38303a3-737e-44cb-a536-6f618e6fa272-logs\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.932228 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-scripts\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.940349 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cbc24b3a-9e3a-4653-8617-45eac69ac198" (UID: "cbc24b3a-9e3a-4653-8617-45eac69ac198"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:33 crc kubenswrapper[4696]: I1125 10:54:33.953255 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s844k\" (UniqueName: \"kubernetes.io/projected/d38303a3-737e-44cb-a536-6f618e6fa272-kube-api-access-s844k\") pod \"horizon-7d6957f485-4l99x\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.000739 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-config" (OuterVolumeSpecName: "config") pod "cbc24b3a-9e3a-4653-8617-45eac69ac198" (UID: "cbc24b3a-9e3a-4653-8617-45eac69ac198"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.011438 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cbc24b3a-9e3a-4653-8617-45eac69ac198" (UID: "cbc24b3a-9e3a-4653-8617-45eac69ac198"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035257 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-config\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035316 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035344 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035382 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-svc\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035427 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zp6m\" (UniqueName: \"kubernetes.io/projected/281499bc-198d-46c6-82e6-ee9260834650-kube-api-access-9zp6m\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035479 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035542 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035552 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.035562 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbc24b3a-9e3a-4653-8617-45eac69ac198-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.038405 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.039215 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-config\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.039544 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-svc\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.046802 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.050073 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.060310 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.069075 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dq4dk" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.114535 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zp6m\" (UniqueName: \"kubernetes.io/projected/281499bc-198d-46c6-82e6-ee9260834650-kube-api-access-9zp6m\") pod \"dnsmasq-dns-cf78879c9-7d95t\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.215029 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.324576 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.343205 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" event={"ID":"cbc24b3a-9e3a-4653-8617-45eac69ac198","Type":"ContainerDied","Data":"1a3e7aa58f78a4235688f6d6cf689c1ce1269b12dde90d0d46948c3b0cdee416"} Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.343264 4696 scope.go:117] "RemoveContainer" containerID="a196b7680557afe9147c63a12d5a58855670014218b4c017d107fdab61427c25" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.343400 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-ljbcg" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.397835 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-ljbcg"] Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.398908 4696 scope.go:117] "RemoveContainer" containerID="50f861584f6abe0a40b4b1df28ec8c6f39277768349ee3db6222df4a73805250" Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.412691 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-ljbcg"] Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.485207 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7557dc88f-7gn6r"] Nov 25 10:54:34 crc kubenswrapper[4696]: W1125 10:54:34.495802 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf437b119_d3e9_4145_9323_56fd522cd092.slice/crio-71617d576e907e595f70e11ec870b5a59c8be693d56ee59a5a1c69a3162af391 WatchSource:0}: Error finding container 71617d576e907e595f70e11ec870b5a59c8be693d56ee59a5a1c69a3162af391: Status 404 returned error can't find the container with id 71617d576e907e595f70e11ec870b5a59c8be693d56ee59a5a1c69a3162af391 Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.815443 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.834188 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5vqmj"] Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.847382 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sbcmj"] Nov 25 10:54:34 crc kubenswrapper[4696]: W1125 10:54:34.872017 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9f225bb_2c7f_4ba8_bd5e_a32a56b0178f.slice/crio-45c8c8ef4241e3f7210a19f910cc3e6ecadc45ea5c11c51dfbce1e5b4d7caca5 WatchSource:0}: Error finding container 45c8c8ef4241e3f7210a19f910cc3e6ecadc45ea5c11c51dfbce1e5b4d7caca5: Status 404 returned error can't find the container with id 45c8c8ef4241e3f7210a19f910cc3e6ecadc45ea5c11c51dfbce1e5b4d7caca5 Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.872981 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8kf4c"] Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.884307 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7qtsq"] Nov 25 10:54:34 crc kubenswrapper[4696]: W1125 10:54:34.891449 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9353d93a_d886_43d0_a4d4_6ad17a28f52a.slice/crio-6e1a110e197d7d13f572214aeb6be9ed52da02dec32060e419454586be90639b WatchSource:0}: Error finding container 6e1a110e197d7d13f572214aeb6be9ed52da02dec32060e419454586be90639b: Status 404 returned error can't find the container with id 6e1a110e197d7d13f572214aeb6be9ed52da02dec32060e419454586be90639b Nov 25 10:54:34 crc kubenswrapper[4696]: I1125 10:54:34.985973 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nqj4p"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.033222 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-7d95t"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.068002 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dq4dk"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.074086 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d6957f485-4l99x"] Nov 25 10:54:35 crc kubenswrapper[4696]: W1125 10:54:35.083254 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a052a79_e780_430c_a049_cc01418ba194.slice/crio-9217b49b3b10ae613e787924163339a7611648d6e4863b20b1f5234a50ed417f WatchSource:0}: Error finding container 9217b49b3b10ae613e787924163339a7611648d6e4863b20b1f5234a50ed417f: Status 404 returned error can't find the container with id 9217b49b3b10ae613e787924163339a7611648d6e4863b20b1f5234a50ed417f Nov 25 10:54:35 crc kubenswrapper[4696]: W1125 10:54:35.136916 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod281499bc_198d_46c6_82e6_ee9260834650.slice/crio-2fb85b74e0b89ab63453ccadbe0efb8532980e1b2359c79c0c09b81777853742 WatchSource:0}: Error finding container 2fb85b74e0b89ab63453ccadbe0efb8532980e1b2359c79c0c09b81777853742: Status 404 returned error can't find the container with id 2fb85b74e0b89ab63453ccadbe0efb8532980e1b2359c79c0c09b81777853742 Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.367122 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" event={"ID":"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f","Type":"ContainerStarted","Data":"45c8c8ef4241e3f7210a19f910cc3e6ecadc45ea5c11c51dfbce1e5b4d7caca5"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.370563 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqj4p" event={"ID":"3a052a79-e780-430c-a049-cc01418ba194","Type":"ContainerStarted","Data":"9217b49b3b10ae613e787924163339a7611648d6e4863b20b1f5234a50ed417f"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.376220 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vqmj" event={"ID":"cacd4fa8-b70a-451d-92a9-40e53ea0a406","Type":"ContainerStarted","Data":"b3cb50826db21c55197adaf101932c8361793a73d6ea060bd00bac195a66b9d3"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.383059 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" event={"ID":"281499bc-198d-46c6-82e6-ee9260834650","Type":"ContainerStarted","Data":"2fb85b74e0b89ab63453ccadbe0efb8532980e1b2359c79c0c09b81777853742"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.385569 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d6957f485-4l99x" event={"ID":"d38303a3-737e-44cb-a536-6f618e6fa272","Type":"ContainerStarted","Data":"b547d830993aadd9626cd1e4bdc0c85b20f499c98bfc3176f60708a30b76def4"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.388539 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7qtsq" event={"ID":"9353d93a-d886-43d0-a4d4-6ad17a28f52a","Type":"ContainerStarted","Data":"6e1a110e197d7d13f572214aeb6be9ed52da02dec32060e419454586be90639b"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.391376 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dq4dk" event={"ID":"c1bfd7e7-4caa-4b79-be66-004376b3f7f4","Type":"ContainerStarted","Data":"7e6242f99070ea873b9371f40e767753f7fb49abd2bc83a6ea336072fb84dd9d"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.411545 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbcmj" event={"ID":"7562e866-74fd-44f7-9d64-b38073fb6fa3","Type":"ContainerStarted","Data":"cf7a90c87b80cd2662b2fa38351695f5036287faefe942eed82dcf39f9c37c7b"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.420763 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7557dc88f-7gn6r" event={"ID":"f437b119-d3e9-4145-9323-56fd522cd092","Type":"ContainerStarted","Data":"71617d576e907e595f70e11ec870b5a59c8be693d56ee59a5a1c69a3162af391"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.423454 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerStarted","Data":"52d57a14c64d62074b66c6acfebe6d94ab2bc51ac094a524ffdee1324b67f7de"} Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.769530 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7557dc88f-7gn6r"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.793124 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.818445 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bf47bd4b5-pk8zr"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.819893 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.887060 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-scripts\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.887127 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e1a5a9-9ea2-446a-bd46-b82699d5081a-logs\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.887165 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd4ws\" (UniqueName: \"kubernetes.io/projected/49e1a5a9-9ea2-446a-bd46-b82699d5081a-kube-api-access-zd4ws\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.887215 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49e1a5a9-9ea2-446a-bd46-b82699d5081a-horizon-secret-key\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.887259 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-config-data\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.901477 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bf47bd4b5-pk8zr"] Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.995127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-scripts\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.995176 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e1a5a9-9ea2-446a-bd46-b82699d5081a-logs\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.995201 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd4ws\" (UniqueName: \"kubernetes.io/projected/49e1a5a9-9ea2-446a-bd46-b82699d5081a-kube-api-access-zd4ws\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.995244 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49e1a5a9-9ea2-446a-bd46-b82699d5081a-horizon-secret-key\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.995277 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-config-data\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.998146 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e1a5a9-9ea2-446a-bd46-b82699d5081a-logs\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.998939 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-scripts\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:35 crc kubenswrapper[4696]: I1125 10:54:35.999083 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-config-data\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.020025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49e1a5a9-9ea2-446a-bd46-b82699d5081a-horizon-secret-key\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.035115 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd4ws\" (UniqueName: \"kubernetes.io/projected/49e1a5a9-9ea2-446a-bd46-b82699d5081a-kube-api-access-zd4ws\") pod \"horizon-5bf47bd4b5-pk8zr\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.108489 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbc24b3a-9e3a-4653-8617-45eac69ac198" path="/var/lib/kubelet/pods/cbc24b3a-9e3a-4653-8617-45eac69ac198/volumes" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.176596 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.475706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbcmj" event={"ID":"7562e866-74fd-44f7-9d64-b38073fb6fa3","Type":"ContainerStarted","Data":"c9bedf8f9d337dffdbf2d4e0c891453886448ec59a81619beb91ae723ecca0dc"} Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.479609 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqj4p" event={"ID":"3a052a79-e780-430c-a049-cc01418ba194","Type":"ContainerStarted","Data":"aea37d144e9450979ed996b37ced91a9f9660ed6c748e718a17f0296d80bb5ad"} Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.483478 4696 generic.go:334] "Generic (PLEG): container finished" podID="281499bc-198d-46c6-82e6-ee9260834650" containerID="7f393b0fbfc79962968f4bf19ff963227a59be385892bae92cbb1d30a4b13350" exitCode=0 Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.483524 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" event={"ID":"281499bc-198d-46c6-82e6-ee9260834650","Type":"ContainerDied","Data":"7f393b0fbfc79962968f4bf19ff963227a59be385892bae92cbb1d30a4b13350"} Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.486765 4696 generic.go:334] "Generic (PLEG): container finished" podID="c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" containerID="a7a871aa7fdba5da88389b9a1148559e20cb50848f7ef5ef3429710ce8055e1f" exitCode=0 Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.486786 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" event={"ID":"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f","Type":"ContainerDied","Data":"a7a871aa7fdba5da88389b9a1148559e20cb50848f7ef5ef3429710ce8055e1f"} Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.503372 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sbcmj" podStartSLOduration=4.503337573 podStartE2EDuration="4.503337573s" podCreationTimestamp="2025-11-25 10:54:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:36.494281855 +0000 UTC m=+1073.307898452" watchObservedRunningTime="2025-11-25 10:54:36.503337573 +0000 UTC m=+1073.316954170" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.571535 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nqj4p" podStartSLOduration=3.5715160089999998 podStartE2EDuration="3.571516009s" podCreationTimestamp="2025-11-25 10:54:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:36.517779187 +0000 UTC m=+1073.331395784" watchObservedRunningTime="2025-11-25 10:54:36.571516009 +0000 UTC m=+1073.385132596" Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.852619 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bf47bd4b5-pk8zr"] Nov 25 10:54:36 crc kubenswrapper[4696]: I1125 10:54:36.909044 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.040140 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq8vb\" (UniqueName: \"kubernetes.io/projected/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-kube-api-access-fq8vb\") pod \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.040571 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-config\") pod \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.040601 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-sb\") pod \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.040629 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-nb\") pod \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.040760 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-svc\") pod \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.040833 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-swift-storage-0\") pod \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\" (UID: \"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f\") " Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.078038 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-kube-api-access-fq8vb" (OuterVolumeSpecName: "kube-api-access-fq8vb") pod "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" (UID: "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f"). InnerVolumeSpecName "kube-api-access-fq8vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.100562 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" (UID: "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.112805 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-config" (OuterVolumeSpecName: "config") pod "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" (UID: "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.143095 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq8vb\" (UniqueName: \"kubernetes.io/projected/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-kube-api-access-fq8vb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.143133 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.143147 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.144535 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" (UID: "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.157767 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" (UID: "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.158273 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" (UID: "c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.246385 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.246416 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.246426 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.535101 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" event={"ID":"281499bc-198d-46c6-82e6-ee9260834650","Type":"ContainerStarted","Data":"83f739bd34ab4041c7712194bde2e25bc9a44af6942bd435438bd50ef7bf6b8e"} Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.535175 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.548898 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" event={"ID":"c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f","Type":"ContainerDied","Data":"45c8c8ef4241e3f7210a19f910cc3e6ecadc45ea5c11c51dfbce1e5b4d7caca5"} Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.548950 4696 scope.go:117] "RemoveContainer" containerID="a7a871aa7fdba5da88389b9a1148559e20cb50848f7ef5ef3429710ce8055e1f" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.549174 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8kf4c" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.561238 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" podStartSLOduration=4.561215188 podStartE2EDuration="4.561215188s" podCreationTimestamp="2025-11-25 10:54:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:37.556766842 +0000 UTC m=+1074.370383439" watchObservedRunningTime="2025-11-25 10:54:37.561215188 +0000 UTC m=+1074.374831765" Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.564262 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bf47bd4b5-pk8zr" event={"ID":"49e1a5a9-9ea2-446a-bd46-b82699d5081a","Type":"ContainerStarted","Data":"c8066416465a9ba14b8f25727f52d33510020b3343ef6239b24f4a7128b25c0a"} Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.645735 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8kf4c"] Nov 25 10:54:37 crc kubenswrapper[4696]: I1125 10:54:37.659620 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8kf4c"] Nov 25 10:54:38 crc kubenswrapper[4696]: I1125 10:54:38.065625 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" path="/var/lib/kubelet/pods/c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f/volumes" Nov 25 10:54:40 crc kubenswrapper[4696]: I1125 10:54:40.604711 4696 generic.go:334] "Generic (PLEG): container finished" podID="bc2593ea-7985-4050-8000-2be38f4b488c" containerID="62be1a71b1f817ac50a79718ab71a3fa8ac78f3bd0933da18172e94b2d922df4" exitCode=0 Nov 25 10:54:40 crc kubenswrapper[4696]: I1125 10:54:40.604806 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ljbtx" event={"ID":"bc2593ea-7985-4050-8000-2be38f4b488c","Type":"ContainerDied","Data":"62be1a71b1f817ac50a79718ab71a3fa8ac78f3bd0933da18172e94b2d922df4"} Nov 25 10:54:41 crc kubenswrapper[4696]: I1125 10:54:41.614027 4696 generic.go:334] "Generic (PLEG): container finished" podID="7562e866-74fd-44f7-9d64-b38073fb6fa3" containerID="c9bedf8f9d337dffdbf2d4e0c891453886448ec59a81619beb91ae723ecca0dc" exitCode=0 Nov 25 10:54:41 crc kubenswrapper[4696]: I1125 10:54:41.614273 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbcmj" event={"ID":"7562e866-74fd-44f7-9d64-b38073fb6fa3","Type":"ContainerDied","Data":"c9bedf8f9d337dffdbf2d4e0c891453886448ec59a81619beb91ae723ecca0dc"} Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.413397 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d6957f485-4l99x"] Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.437637 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c8fd89c8d-5vb5f"] Nov 25 10:54:42 crc kubenswrapper[4696]: E1125 10:54:42.438173 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" containerName="init" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.438201 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" containerName="init" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.438433 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f225bb-2c7f-4ba8-bd5e-a32a56b0178f" containerName="init" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.439614 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.443707 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468471 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h88gp\" (UniqueName: \"kubernetes.io/projected/28158e14-5724-410c-9d4c-5ca5563f6f87-kube-api-access-h88gp\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468518 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-tls-certs\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468567 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-config-data\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468591 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-scripts\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468657 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-secret-key\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468712 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28158e14-5724-410c-9d4c-5ca5563f6f87-logs\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.468779 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-combined-ca-bundle\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.471459 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c8fd89c8d-5vb5f"] Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.521721 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bf47bd4b5-pk8zr"] Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.550903 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6558fc797b-2qp4r"] Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.552320 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.569935 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-combined-ca-bundle\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570047 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h88gp\" (UniqueName: \"kubernetes.io/projected/28158e14-5724-410c-9d4c-5ca5563f6f87-kube-api-access-h88gp\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570062 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6558fc797b-2qp4r"] Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570073 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-tls-certs\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570110 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-config-data\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570129 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-scripts\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570181 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-secret-key\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570207 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28158e14-5724-410c-9d4c-5ca5563f6f87-logs\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.570713 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28158e14-5724-410c-9d4c-5ca5563f6f87-logs\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.573203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-scripts\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.573865 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-config-data\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.599064 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-secret-key\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.600022 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-tls-certs\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.604678 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-combined-ca-bundle\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.608384 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h88gp\" (UniqueName: \"kubernetes.io/projected/28158e14-5724-410c-9d4c-5ca5563f6f87-kube-api-access-h88gp\") pod \"horizon-7c8fd89c8d-5vb5f\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672002 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1944527b-6ffc-4b6b-954f-4b01394ea0cf-logs\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672254 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-combined-ca-bundle\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672331 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-horizon-tls-certs\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672387 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1944527b-6ffc-4b6b-954f-4b01394ea0cf-config-data\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672414 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1944527b-6ffc-4b6b-954f-4b01394ea0cf-scripts\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672430 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-horizon-secret-key\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.672450 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz9rh\" (UniqueName: \"kubernetes.io/projected/1944527b-6ffc-4b6b-954f-4b01394ea0cf-kube-api-access-lz9rh\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.759934 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777295 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-combined-ca-bundle\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-horizon-tls-certs\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777460 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1944527b-6ffc-4b6b-954f-4b01394ea0cf-config-data\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777492 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1944527b-6ffc-4b6b-954f-4b01394ea0cf-scripts\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777516 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-horizon-secret-key\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777543 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz9rh\" (UniqueName: \"kubernetes.io/projected/1944527b-6ffc-4b6b-954f-4b01394ea0cf-kube-api-access-lz9rh\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.777592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1944527b-6ffc-4b6b-954f-4b01394ea0cf-logs\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.778119 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1944527b-6ffc-4b6b-954f-4b01394ea0cf-logs\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.780605 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1944527b-6ffc-4b6b-954f-4b01394ea0cf-scripts\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.782190 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1944527b-6ffc-4b6b-954f-4b01394ea0cf-config-data\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.783534 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-combined-ca-bundle\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.787279 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-horizon-secret-key\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.791404 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1944527b-6ffc-4b6b-954f-4b01394ea0cf-horizon-tls-certs\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.803135 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz9rh\" (UniqueName: \"kubernetes.io/projected/1944527b-6ffc-4b6b-954f-4b01394ea0cf-kube-api-access-lz9rh\") pod \"horizon-6558fc797b-2qp4r\" (UID: \"1944527b-6ffc-4b6b-954f-4b01394ea0cf\") " pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:42 crc kubenswrapper[4696]: I1125 10:54:42.875565 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:54:44 crc kubenswrapper[4696]: I1125 10:54:44.326879 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:54:44 crc kubenswrapper[4696]: I1125 10:54:44.394861 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-wtvbr"] Nov 25 10:54:44 crc kubenswrapper[4696]: I1125 10:54:44.395128 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" containerID="cri-o://8f50f81db271f4f47fc5b1a6901dce23cc6399e8c12abe5901b52750df59aa40" gracePeriod=10 Nov 25 10:54:44 crc kubenswrapper[4696]: I1125 10:54:44.664369 4696 generic.go:334] "Generic (PLEG): container finished" podID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerID="8f50f81db271f4f47fc5b1a6901dce23cc6399e8c12abe5901b52750df59aa40" exitCode=0 Nov 25 10:54:44 crc kubenswrapper[4696]: I1125 10:54:44.664453 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" event={"ID":"6aac8f88-5b55-4f25-8cee-6b20fa9faecd","Type":"ContainerDied","Data":"8f50f81db271f4f47fc5b1a6901dce23cc6399e8c12abe5901b52750df59aa40"} Nov 25 10:54:46 crc kubenswrapper[4696]: I1125 10:54:46.812733 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Nov 25 10:54:51 crc kubenswrapper[4696]: I1125 10:54:51.813176 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Nov 25 10:54:55 crc kubenswrapper[4696]: E1125 10:54:55.386305 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 25 10:54:55 crc kubenswrapper[4696]: E1125 10:54:55.386970 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n645h6bh694h84h694h8ch54ch55bh597h657h5c9hd9h66ch658h657h597hcfhb7h5b9hc7h64h54fh594hd6h68h598h6bhf5h55bh644h68bh5f8q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hbqtl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7557dc88f-7gn6r_openstack(f437b119-d3e9-4145-9323-56fd522cd092): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:54:55 crc kubenswrapper[4696]: E1125 10:54:55.395361 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7557dc88f-7gn6r" podUID="f437b119-d3e9-4145-9323-56fd522cd092" Nov 25 10:54:55 crc kubenswrapper[4696]: E1125 10:54:55.420345 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 25 10:54:55 crc kubenswrapper[4696]: E1125 10:54:55.420508 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n696h7h5b7h5c9h675h686h64fh54bh55h659h596h5fdh5fch8bh6dh64bhcdh584h8ch5b7h89h5dch664hb9h7dh568h5c9h5dch8fhbch65hd9q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zd4ws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5bf47bd4b5-pk8zr_openstack(49e1a5a9-9ea2-446a-bd46-b82699d5081a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:54:55 crc kubenswrapper[4696]: E1125 10:54:55.422506 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5bf47bd4b5-pk8zr" podUID="49e1a5a9-9ea2-446a-bd46-b82699d5081a" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.487812 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ljbtx" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.523116 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-combined-ca-bundle\") pod \"bc2593ea-7985-4050-8000-2be38f4b488c\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.523218 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-config-data\") pod \"bc2593ea-7985-4050-8000-2be38f4b488c\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.523253 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-db-sync-config-data\") pod \"bc2593ea-7985-4050-8000-2be38f4b488c\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.523293 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqkgx\" (UniqueName: \"kubernetes.io/projected/bc2593ea-7985-4050-8000-2be38f4b488c-kube-api-access-qqkgx\") pod \"bc2593ea-7985-4050-8000-2be38f4b488c\" (UID: \"bc2593ea-7985-4050-8000-2be38f4b488c\") " Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.529303 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2593ea-7985-4050-8000-2be38f4b488c-kube-api-access-qqkgx" (OuterVolumeSpecName: "kube-api-access-qqkgx") pod "bc2593ea-7985-4050-8000-2be38f4b488c" (UID: "bc2593ea-7985-4050-8000-2be38f4b488c"). InnerVolumeSpecName "kube-api-access-qqkgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.558985 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bc2593ea-7985-4050-8000-2be38f4b488c" (UID: "bc2593ea-7985-4050-8000-2be38f4b488c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.579059 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc2593ea-7985-4050-8000-2be38f4b488c" (UID: "bc2593ea-7985-4050-8000-2be38f4b488c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.620752 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-config-data" (OuterVolumeSpecName: "config-data") pod "bc2593ea-7985-4050-8000-2be38f4b488c" (UID: "bc2593ea-7985-4050-8000-2be38f4b488c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.625871 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.625894 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.625917 4696 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2593ea-7985-4050-8000-2be38f4b488c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.625925 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqkgx\" (UniqueName: \"kubernetes.io/projected/bc2593ea-7985-4050-8000-2be38f4b488c-kube-api-access-qqkgx\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.782479 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ljbtx" Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.786728 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ljbtx" event={"ID":"bc2593ea-7985-4050-8000-2be38f4b488c","Type":"ContainerDied","Data":"b9823b1a9f41cb38b25426e3eda10ea6bd4d5940999cf67a2e6e0fdcba4d79ae"} Nov 25 10:54:55 crc kubenswrapper[4696]: I1125 10:54:55.787584 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9823b1a9f41cb38b25426e3eda10ea6bd4d5940999cf67a2e6e0fdcba4d79ae" Nov 25 10:54:56 crc kubenswrapper[4696]: I1125 10:54:56.991343 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-f8bqt"] Nov 25 10:54:56 crc kubenswrapper[4696]: E1125 10:54:56.991952 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2593ea-7985-4050-8000-2be38f4b488c" containerName="glance-db-sync" Nov 25 10:54:56 crc kubenswrapper[4696]: I1125 10:54:56.991965 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2593ea-7985-4050-8000-2be38f4b488c" containerName="glance-db-sync" Nov 25 10:54:56 crc kubenswrapper[4696]: I1125 10:54:56.992146 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2593ea-7985-4050-8000-2be38f4b488c" containerName="glance-db-sync" Nov 25 10:54:56 crc kubenswrapper[4696]: I1125 10:54:56.993009 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.016220 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-f8bqt"] Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.060846 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.060996 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65pnx\" (UniqueName: \"kubernetes.io/projected/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-kube-api-access-65pnx\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.061074 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.061124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-config\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.061187 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.061263 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.162719 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.162827 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65pnx\" (UniqueName: \"kubernetes.io/projected/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-kube-api-access-65pnx\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.162864 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.162894 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-config\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.162927 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.162957 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.163800 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.163828 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.164027 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.164300 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-config\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.164737 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.193981 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65pnx\" (UniqueName: \"kubernetes.io/projected/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-kube-api-access-65pnx\") pod \"dnsmasq-dns-56df8fb6b7-f8bqt\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.323985 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:54:57 crc kubenswrapper[4696]: E1125 10:54:57.837285 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 25 10:54:57 crc kubenswrapper[4696]: E1125 10:54:57.837454 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d2c2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-dq4dk_openstack(c1bfd7e7-4caa-4b79-be66-004376b3f7f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:54:57 crc kubenswrapper[4696]: E1125 10:54:57.839935 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-dq4dk" podUID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.861917 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.863637 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.869351 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.869605 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.869811 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-s5c2l" Nov 25 10:54:57 crc kubenswrapper[4696]: E1125 10:54:57.876889 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 25 10:54:57 crc kubenswrapper[4696]: E1125 10:54:57.877043 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbfhcbh58bh64fh7dh5bbh64fh59fh9dh55bhc9h59bh57ch79h67ch5cbh55bh567h56bh5chbch5dbh56chd8hcdh7dh567h68bh78h694h678h55bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s844k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d6957f485-4l99x_openstack(d38303a3-737e-44cb-a536-6f618e6fa272): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:54:57 crc kubenswrapper[4696]: E1125 10:54:57.880791 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d6957f485-4l99x" podUID="d38303a3-737e-44cb-a536-6f618e6fa272" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.890873 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.929440 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.979529 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht79p\" (UniqueName: \"kubernetes.io/projected/7562e866-74fd-44f7-9d64-b38073fb6fa3-kube-api-access-ht79p\") pod \"7562e866-74fd-44f7-9d64-b38073fb6fa3\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.979632 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-fernet-keys\") pod \"7562e866-74fd-44f7-9d64-b38073fb6fa3\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.979767 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-combined-ca-bundle\") pod \"7562e866-74fd-44f7-9d64-b38073fb6fa3\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.979929 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-config-data\") pod \"7562e866-74fd-44f7-9d64-b38073fb6fa3\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.979966 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-scripts\") pod \"7562e866-74fd-44f7-9d64-b38073fb6fa3\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.979998 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-credential-keys\") pod \"7562e866-74fd-44f7-9d64-b38073fb6fa3\" (UID: \"7562e866-74fd-44f7-9d64-b38073fb6fa3\") " Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980243 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980285 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980318 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-logs\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980365 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqhg6\" (UniqueName: \"kubernetes.io/projected/cc1c6f7a-035e-420f-a056-f63d0c8babcd-kube-api-access-nqhg6\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980416 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980437 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.980509 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.986010 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7562e866-74fd-44f7-9d64-b38073fb6fa3" (UID: "7562e866-74fd-44f7-9d64-b38073fb6fa3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.987152 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7562e866-74fd-44f7-9d64-b38073fb6fa3-kube-api-access-ht79p" (OuterVolumeSpecName: "kube-api-access-ht79p") pod "7562e866-74fd-44f7-9d64-b38073fb6fa3" (UID: "7562e866-74fd-44f7-9d64-b38073fb6fa3"). InnerVolumeSpecName "kube-api-access-ht79p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.988342 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7562e866-74fd-44f7-9d64-b38073fb6fa3" (UID: "7562e866-74fd-44f7-9d64-b38073fb6fa3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4696]: I1125 10:54:57.995923 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-scripts" (OuterVolumeSpecName: "scripts") pod "7562e866-74fd-44f7-9d64-b38073fb6fa3" (UID: "7562e866-74fd-44f7-9d64-b38073fb6fa3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.034851 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7562e866-74fd-44f7-9d64-b38073fb6fa3" (UID: "7562e866-74fd-44f7-9d64-b38073fb6fa3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.056231 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-config-data" (OuterVolumeSpecName: "config-data") pod "7562e866-74fd-44f7-9d64-b38073fb6fa3" (UID: "7562e866-74fd-44f7-9d64-b38073fb6fa3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086042 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086116 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086143 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-logs\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086191 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqhg6\" (UniqueName: \"kubernetes.io/projected/cc1c6f7a-035e-420f-a056-f63d0c8babcd-kube-api-access-nqhg6\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086235 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086259 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086357 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086443 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086460 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086468 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086478 4696 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086488 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht79p\" (UniqueName: \"kubernetes.io/projected/7562e866-74fd-44f7-9d64-b38073fb6fa3-kube-api-access-ht79p\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.086497 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7562e866-74fd-44f7-9d64-b38073fb6fa3-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.088856 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.090062 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-logs\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.090330 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.094590 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.094881 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.095256 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.115625 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqhg6\" (UniqueName: \"kubernetes.io/projected/cc1c6f7a-035e-420f-a056-f63d0c8babcd-kube-api-access-nqhg6\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.123290 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.212102 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:54:58 crc kubenswrapper[4696]: E1125 10:54:58.212607 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7562e866-74fd-44f7-9d64-b38073fb6fa3" containerName="keystone-bootstrap" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.212647 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7562e866-74fd-44f7-9d64-b38073fb6fa3" containerName="keystone-bootstrap" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.213023 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7562e866-74fd-44f7-9d64-b38073fb6fa3" containerName="keystone-bootstrap" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.214265 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.216857 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.230740 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.246369 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.289629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.289793 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.289828 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwm2f\" (UniqueName: \"kubernetes.io/projected/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-kube-api-access-bwm2f\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.289857 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.290024 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.290043 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.290092 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392035 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392117 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392182 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392224 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392264 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392276 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392286 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwm2f\" (UniqueName: \"kubernetes.io/projected/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-kube-api-access-bwm2f\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.392934 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.393639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.397593 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.411582 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.412207 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.429175 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwm2f\" (UniqueName: \"kubernetes.io/projected/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-kube-api-access-bwm2f\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.441432 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.569729 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.816142 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbcmj" event={"ID":"7562e866-74fd-44f7-9d64-b38073fb6fa3","Type":"ContainerDied","Data":"cf7a90c87b80cd2662b2fa38351695f5036287faefe942eed82dcf39f9c37c7b"} Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.816190 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf7a90c87b80cd2662b2fa38351695f5036287faefe942eed82dcf39f9c37c7b" Nov 25 10:54:58 crc kubenswrapper[4696]: I1125 10:54:58.816240 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbcmj" Nov 25 10:54:58 crc kubenswrapper[4696]: E1125 10:54:58.820428 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-dq4dk" podUID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.103854 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sbcmj"] Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.120099 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sbcmj"] Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.210288 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-c2g2r"] Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.211787 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.217778 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.217948 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.218079 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7vqt" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.218175 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.218282 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.223930 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c2g2r"] Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.312172 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-credential-keys\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.312211 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-scripts\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.312236 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-combined-ca-bundle\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.312264 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-config-data\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.312294 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5bhh\" (UniqueName: \"kubernetes.io/projected/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-kube-api-access-b5bhh\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.312312 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-fernet-keys\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.412825 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-credential-keys\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.412856 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-scripts\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.412883 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-combined-ca-bundle\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.412912 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-config-data\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.412943 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5bhh\" (UniqueName: \"kubernetes.io/projected/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-kube-api-access-b5bhh\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.412962 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-fernet-keys\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.420585 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-config-data\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.421739 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-scripts\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.422190 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-fernet-keys\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.422205 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-credential-keys\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.430757 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-combined-ca-bundle\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.436620 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5bhh\" (UniqueName: \"kubernetes.io/projected/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-kube-api-access-b5bhh\") pod \"keystone-bootstrap-c2g2r\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.529420 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.860810 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:54:59 crc kubenswrapper[4696]: I1125 10:54:59.930219 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:00 crc kubenswrapper[4696]: I1125 10:55:00.058555 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7562e866-74fd-44f7-9d64-b38073fb6fa3" path="/var/lib/kubelet/pods/7562e866-74fd-44f7-9d64-b38073fb6fa3/volumes" Nov 25 10:55:01 crc kubenswrapper[4696]: I1125 10:55:01.815089 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 25 10:55:01 crc kubenswrapper[4696]: I1125 10:55:01.817502 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:55:06 crc kubenswrapper[4696]: I1125 10:55:06.815287 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 25 10:55:11 crc kubenswrapper[4696]: I1125 10:55:11.816763 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 25 10:55:16 crc kubenswrapper[4696]: I1125 10:55:16.818006 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 25 10:55:21 crc kubenswrapper[4696]: I1125 10:55:21.819805 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.012822 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bf47bd4b5-pk8zr" event={"ID":"49e1a5a9-9ea2-446a-bd46-b82699d5081a","Type":"ContainerDied","Data":"c8066416465a9ba14b8f25727f52d33510020b3343ef6239b24f4a7128b25c0a"} Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.012864 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8066416465a9ba14b8f25727f52d33510020b3343ef6239b24f4a7128b25c0a" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.015840 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" event={"ID":"6aac8f88-5b55-4f25-8cee-6b20fa9faecd","Type":"ContainerDied","Data":"76d5c07cd580b6165ef07a66408ed480c7835e40e5b1d22e3c23fafc645978ee"} Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.016272 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76d5c07cd580b6165ef07a66408ed480c7835e40e5b1d22e3c23fafc645978ee" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.017897 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d6957f485-4l99x" event={"ID":"d38303a3-737e-44cb-a536-6f618e6fa272","Type":"ContainerDied","Data":"b547d830993aadd9626cd1e4bdc0c85b20f499c98bfc3176f60708a30b76def4"} Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.017954 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b547d830993aadd9626cd1e4bdc0c85b20f499c98bfc3176f60708a30b76def4" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.019632 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7557dc88f-7gn6r" event={"ID":"f437b119-d3e9-4145-9323-56fd522cd092","Type":"ContainerDied","Data":"71617d576e907e595f70e11ec870b5a59c8be693d56ee59a5a1c69a3162af391"} Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.019657 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71617d576e907e595f70e11ec870b5a59c8be693d56ee59a5a1c69a3162af391" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.030825 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.036761 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.042983 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.068629 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215211 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-scripts\") pod \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215302 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-config-data\") pod \"f437b119-d3e9-4145-9323-56fd522cd092\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215344 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f437b119-d3e9-4145-9323-56fd522cd092-horizon-secret-key\") pod \"f437b119-d3e9-4145-9323-56fd522cd092\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215381 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38303a3-737e-44cb-a536-6f618e6fa272-logs\") pod \"d38303a3-737e-44cb-a536-6f618e6fa272\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215423 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f437b119-d3e9-4145-9323-56fd522cd092-logs\") pod \"f437b119-d3e9-4145-9323-56fd522cd092\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215463 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-config-data\") pod \"d38303a3-737e-44cb-a536-6f618e6fa272\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215526 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd4ws\" (UniqueName: \"kubernetes.io/projected/49e1a5a9-9ea2-446a-bd46-b82699d5081a-kube-api-access-zd4ws\") pod \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215583 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbqtl\" (UniqueName: \"kubernetes.io/projected/f437b119-d3e9-4145-9323-56fd522cd092-kube-api-access-hbqtl\") pod \"f437b119-d3e9-4145-9323-56fd522cd092\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215620 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-config-data\") pod \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215710 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-svc\") pod \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215766 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-nb\") pod \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-sb\") pod \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215853 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhk6h\" (UniqueName: \"kubernetes.io/projected/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-kube-api-access-zhk6h\") pod \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215896 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s844k\" (UniqueName: \"kubernetes.io/projected/d38303a3-737e-44cb-a536-6f618e6fa272-kube-api-access-s844k\") pod \"d38303a3-737e-44cb-a536-6f618e6fa272\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215887 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-scripts" (OuterVolumeSpecName: "scripts") pod "49e1a5a9-9ea2-446a-bd46-b82699d5081a" (UID: "49e1a5a9-9ea2-446a-bd46-b82699d5081a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215941 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-swift-storage-0\") pod \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.215990 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-scripts\") pod \"d38303a3-737e-44cb-a536-6f618e6fa272\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216024 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-scripts\") pod \"f437b119-d3e9-4145-9323-56fd522cd092\" (UID: \"f437b119-d3e9-4145-9323-56fd522cd092\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216089 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d38303a3-737e-44cb-a536-6f618e6fa272-horizon-secret-key\") pod \"d38303a3-737e-44cb-a536-6f618e6fa272\" (UID: \"d38303a3-737e-44cb-a536-6f618e6fa272\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216126 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-config\") pod \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\" (UID: \"6aac8f88-5b55-4f25-8cee-6b20fa9faecd\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216164 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49e1a5a9-9ea2-446a-bd46-b82699d5081a-horizon-secret-key\") pod \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216223 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e1a5a9-9ea2-446a-bd46-b82699d5081a-logs\") pod \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\" (UID: \"49e1a5a9-9ea2-446a-bd46-b82699d5081a\") " Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216304 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-config-data" (OuterVolumeSpecName: "config-data") pod "f437b119-d3e9-4145-9323-56fd522cd092" (UID: "f437b119-d3e9-4145-9323-56fd522cd092"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.216878 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-config-data" (OuterVolumeSpecName: "config-data") pod "d38303a3-737e-44cb-a536-6f618e6fa272" (UID: "d38303a3-737e-44cb-a536-6f618e6fa272"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.219827 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-scripts" (OuterVolumeSpecName: "scripts") pod "d38303a3-737e-44cb-a536-6f618e6fa272" (UID: "d38303a3-737e-44cb-a536-6f618e6fa272"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.222170 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-scripts" (OuterVolumeSpecName: "scripts") pod "f437b119-d3e9-4145-9323-56fd522cd092" (UID: "f437b119-d3e9-4145-9323-56fd522cd092"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.222362 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-config-data" (OuterVolumeSpecName: "config-data") pod "49e1a5a9-9ea2-446a-bd46-b82699d5081a" (UID: "49e1a5a9-9ea2-446a-bd46-b82699d5081a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.222479 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e1a5a9-9ea2-446a-bd46-b82699d5081a-kube-api-access-zd4ws" (OuterVolumeSpecName: "kube-api-access-zd4ws") pod "49e1a5a9-9ea2-446a-bd46-b82699d5081a" (UID: "49e1a5a9-9ea2-446a-bd46-b82699d5081a"). InnerVolumeSpecName "kube-api-access-zd4ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.217139 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.223062 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.223464 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e1a5a9-9ea2-446a-bd46-b82699d5081a-logs" (OuterVolumeSpecName: "logs") pod "49e1a5a9-9ea2-446a-bd46-b82699d5081a" (UID: "49e1a5a9-9ea2-446a-bd46-b82699d5081a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.223998 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f437b119-d3e9-4145-9323-56fd522cd092-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f437b119-d3e9-4145-9323-56fd522cd092" (UID: "f437b119-d3e9-4145-9323-56fd522cd092"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.226698 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f437b119-d3e9-4145-9323-56fd522cd092-kube-api-access-hbqtl" (OuterVolumeSpecName: "kube-api-access-hbqtl") pod "f437b119-d3e9-4145-9323-56fd522cd092" (UID: "f437b119-d3e9-4145-9323-56fd522cd092"). InnerVolumeSpecName "kube-api-access-hbqtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.227422 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e1a5a9-9ea2-446a-bd46-b82699d5081a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "49e1a5a9-9ea2-446a-bd46-b82699d5081a" (UID: "49e1a5a9-9ea2-446a-bd46-b82699d5081a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.228263 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38303a3-737e-44cb-a536-6f618e6fa272-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d38303a3-737e-44cb-a536-6f618e6fa272" (UID: "d38303a3-737e-44cb-a536-6f618e6fa272"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.228579 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f437b119-d3e9-4145-9323-56fd522cd092-logs" (OuterVolumeSpecName: "logs") pod "f437b119-d3e9-4145-9323-56fd522cd092" (UID: "f437b119-d3e9-4145-9323-56fd522cd092"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.229763 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38303a3-737e-44cb-a536-6f618e6fa272-logs" (OuterVolumeSpecName: "logs") pod "d38303a3-737e-44cb-a536-6f618e6fa272" (UID: "d38303a3-737e-44cb-a536-6f618e6fa272"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.230698 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38303a3-737e-44cb-a536-6f618e6fa272-kube-api-access-s844k" (OuterVolumeSpecName: "kube-api-access-s844k") pod "d38303a3-737e-44cb-a536-6f618e6fa272" (UID: "d38303a3-737e-44cb-a536-6f618e6fa272"). InnerVolumeSpecName "kube-api-access-s844k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.233012 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-kube-api-access-zhk6h" (OuterVolumeSpecName: "kube-api-access-zhk6h") pod "6aac8f88-5b55-4f25-8cee-6b20fa9faecd" (UID: "6aac8f88-5b55-4f25-8cee-6b20fa9faecd"). InnerVolumeSpecName "kube-api-access-zhk6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.267840 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6aac8f88-5b55-4f25-8cee-6b20fa9faecd" (UID: "6aac8f88-5b55-4f25-8cee-6b20fa9faecd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.274751 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6aac8f88-5b55-4f25-8cee-6b20fa9faecd" (UID: "6aac8f88-5b55-4f25-8cee-6b20fa9faecd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.276137 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6aac8f88-5b55-4f25-8cee-6b20fa9faecd" (UID: "6aac8f88-5b55-4f25-8cee-6b20fa9faecd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.297009 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6aac8f88-5b55-4f25-8cee-6b20fa9faecd" (UID: "6aac8f88-5b55-4f25-8cee-6b20fa9faecd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.299106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-config" (OuterVolumeSpecName: "config") pod "6aac8f88-5b55-4f25-8cee-6b20fa9faecd" (UID: "6aac8f88-5b55-4f25-8cee-6b20fa9faecd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323879 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323922 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323934 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f437b119-d3e9-4145-9323-56fd522cd092-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323948 4696 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d38303a3-737e-44cb-a536-6f618e6fa272-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323961 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323973 4696 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49e1a5a9-9ea2-446a-bd46-b82699d5081a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323984 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e1a5a9-9ea2-446a-bd46-b82699d5081a-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.323997 4696 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f437b119-d3e9-4145-9323-56fd522cd092-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324009 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38303a3-737e-44cb-a536-6f618e6fa272-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324020 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f437b119-d3e9-4145-9323-56fd522cd092-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324032 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38303a3-737e-44cb-a536-6f618e6fa272-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324044 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd4ws\" (UniqueName: \"kubernetes.io/projected/49e1a5a9-9ea2-446a-bd46-b82699d5081a-kube-api-access-zd4ws\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324057 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbqtl\" (UniqueName: \"kubernetes.io/projected/f437b119-d3e9-4145-9323-56fd522cd092-kube-api-access-hbqtl\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324068 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49e1a5a9-9ea2-446a-bd46-b82699d5081a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324080 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324091 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324105 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324117 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhk6h\" (UniqueName: \"kubernetes.io/projected/6aac8f88-5b55-4f25-8cee-6b20fa9faecd-kube-api-access-zhk6h\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:22 crc kubenswrapper[4696]: I1125 10:55:22.324130 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s844k\" (UniqueName: \"kubernetes.io/projected/d38303a3-737e-44cb-a536-6f618e6fa272-kube-api-access-s844k\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.026287 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bf47bd4b5-pk8zr" Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.026582 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.026552 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7557dc88f-7gn6r" Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.026514 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d6957f485-4l99x" Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.102715 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-wtvbr"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.124318 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-wtvbr"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.182833 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7557dc88f-7gn6r"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.191503 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7557dc88f-7gn6r"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.279324 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d6957f485-4l99x"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.308052 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d6957f485-4l99x"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.331966 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bf47bd4b5-pk8zr"] Nov 25 10:55:23 crc kubenswrapper[4696]: I1125 10:55:23.338509 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bf47bd4b5-pk8zr"] Nov 25 10:55:24 crc kubenswrapper[4696]: I1125 10:55:24.058602 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e1a5a9-9ea2-446a-bd46-b82699d5081a" path="/var/lib/kubelet/pods/49e1a5a9-9ea2-446a-bd46-b82699d5081a/volumes" Nov 25 10:55:24 crc kubenswrapper[4696]: I1125 10:55:24.059221 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" path="/var/lib/kubelet/pods/6aac8f88-5b55-4f25-8cee-6b20fa9faecd/volumes" Nov 25 10:55:24 crc kubenswrapper[4696]: I1125 10:55:24.059947 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38303a3-737e-44cb-a536-6f618e6fa272" path="/var/lib/kubelet/pods/d38303a3-737e-44cb-a536-6f618e6fa272/volumes" Nov 25 10:55:24 crc kubenswrapper[4696]: I1125 10:55:24.060473 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f437b119-d3e9-4145-9323-56fd522cd092" path="/var/lib/kubelet/pods/f437b119-d3e9-4145-9323-56fd522cd092/volumes" Nov 25 10:55:26 crc kubenswrapper[4696]: E1125 10:55:26.676217 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 25 10:55:26 crc kubenswrapper[4696]: E1125 10:55:26.677550 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k5xg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-5vqmj_openstack(cacd4fa8-b70a-451d-92a9-40e53ea0a406): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:55:26 crc kubenswrapper[4696]: E1125 10:55:26.678987 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-5vqmj" podUID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" Nov 25 10:55:26 crc kubenswrapper[4696]: I1125 10:55:26.821510 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-wtvbr" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 25 10:55:27 crc kubenswrapper[4696]: E1125 10:55:27.066153 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-5vqmj" podUID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" Nov 25 10:55:28 crc kubenswrapper[4696]: E1125 10:55:28.223456 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 25 10:55:28 crc kubenswrapper[4696]: E1125 10:55:28.223874 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-skp7d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7qtsq_openstack(9353d93a-d886-43d0-a4d4-6ad17a28f52a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:55:28 crc kubenswrapper[4696]: E1125 10:55:28.225951 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7qtsq" podUID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" Nov 25 10:55:28 crc kubenswrapper[4696]: I1125 10:55:28.862777 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c8fd89c8d-5vb5f"] Nov 25 10:55:28 crc kubenswrapper[4696]: W1125 10:55:28.865423 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28158e14_5724_410c_9d4c_5ca5563f6f87.slice/crio-2e7adca82ed91f7e8f4bf14ec40ea94cb01cac7b98d14a1bc9e48e4198bd2743 WatchSource:0}: Error finding container 2e7adca82ed91f7e8f4bf14ec40ea94cb01cac7b98d14a1bc9e48e4198bd2743: Status 404 returned error can't find the container with id 2e7adca82ed91f7e8f4bf14ec40ea94cb01cac7b98d14a1bc9e48e4198bd2743 Nov 25 10:55:28 crc kubenswrapper[4696]: I1125 10:55:28.960109 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6558fc797b-2qp4r"] Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.056727 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:55:29 crc kubenswrapper[4696]: W1125 10:55:29.057036 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc1c6f7a_035e_420f_a056_f63d0c8babcd.slice/crio-0eb998a38e8d5cfa4964f82d4355fde07b5ccb94e6d6bd6c1b727f8a59d72894 WatchSource:0}: Error finding container 0eb998a38e8d5cfa4964f82d4355fde07b5ccb94e6d6bd6c1b727f8a59d72894: Status 404 returned error can't find the container with id 0eb998a38e8d5cfa4964f82d4355fde07b5ccb94e6d6bd6c1b727f8a59d72894 Nov 25 10:55:29 crc kubenswrapper[4696]: W1125 10:55:29.073174 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c3a8e43_bb23_4fb2_ab40_e2ff6676212a.slice/crio-d39ddd8d56175f02d7ef3b7fae3609b2f725339d44ab57f250f0577009470f4a WatchSource:0}: Error finding container d39ddd8d56175f02d7ef3b7fae3609b2f725339d44ab57f250f0577009470f4a: Status 404 returned error can't find the container with id d39ddd8d56175f02d7ef3b7fae3609b2f725339d44ab57f250f0577009470f4a Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.095793 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6558fc797b-2qp4r" event={"ID":"1944527b-6ffc-4b6b-954f-4b01394ea0cf","Type":"ContainerStarted","Data":"228b628e91497f70e019e398dc6b2e583a02c42d1d684501c12c25fbb4c99638"} Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.098708 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-f8bqt"] Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.110081 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerStarted","Data":"2e7adca82ed91f7e8f4bf14ec40ea94cb01cac7b98d14a1bc9e48e4198bd2743"} Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.111866 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc1c6f7a-035e-420f-a056-f63d0c8babcd","Type":"ContainerStarted","Data":"0eb998a38e8d5cfa4964f82d4355fde07b5ccb94e6d6bd6c1b727f8a59d72894"} Nov 25 10:55:29 crc kubenswrapper[4696]: W1125 10:55:29.113158 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5c18c5a_6cef_4afe_a3a1_f0ad04cec1a6.slice/crio-3841be5a59d048e0af33a74ae262799efaaf64a36364282614578b569b6d53c2 WatchSource:0}: Error finding container 3841be5a59d048e0af33a74ae262799efaaf64a36364282614578b569b6d53c2: Status 404 returned error can't find the container with id 3841be5a59d048e0af33a74ae262799efaaf64a36364282614578b569b6d53c2 Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.113841 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerStarted","Data":"b129008fc63a532aaeecc84712dd6d46fb99265b4c627542ad08d8514e68301d"} Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.117047 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dq4dk" event={"ID":"c1bfd7e7-4caa-4b79-be66-004376b3f7f4","Type":"ContainerStarted","Data":"5c97bb9514113109c93a46209e8213df0786b15a142aff872fbb81cc5256e3a0"} Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.119384 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" event={"ID":"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a","Type":"ContainerStarted","Data":"d39ddd8d56175f02d7ef3b7fae3609b2f725339d44ab57f250f0577009470f4a"} Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.120113 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c2g2r"] Nov 25 10:55:29 crc kubenswrapper[4696]: E1125 10:55:29.120220 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7qtsq" podUID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.145501 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dq4dk" podStartSLOduration=2.927523131 podStartE2EDuration="56.145483477s" podCreationTimestamp="2025-11-25 10:54:33 +0000 UTC" firstStartedPulling="2025-11-25 10:54:35.092832341 +0000 UTC m=+1071.906448928" lastFinishedPulling="2025-11-25 10:55:28.310792687 +0000 UTC m=+1125.124409274" observedRunningTime="2025-11-25 10:55:29.143125693 +0000 UTC m=+1125.956742300" watchObservedRunningTime="2025-11-25 10:55:29.145483477 +0000 UTC m=+1125.959100074" Nov 25 10:55:29 crc kubenswrapper[4696]: I1125 10:55:29.182266 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.175118 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerStarted","Data":"2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.177916 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc1c6f7a-035e-420f-a056-f63d0c8babcd","Type":"ContainerStarted","Data":"3596bf5ef27d2a873057fc6aab0e0028ec618e63f6bf238d827b811342547d38"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.178962 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9","Type":"ContainerStarted","Data":"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.178987 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9","Type":"ContainerStarted","Data":"95b71d0f28bde65752d6a6d7bccf95dce466324ed92b11b440984dda4947b6fc"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.180176 4696 generic.go:334] "Generic (PLEG): container finished" podID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerID="27c33a1491b2001d59ab6a25a63b79c0bc5160c423bf405f0d1b1bd59a049496" exitCode=0 Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.180217 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" event={"ID":"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a","Type":"ContainerDied","Data":"27c33a1491b2001d59ab6a25a63b79c0bc5160c423bf405f0d1b1bd59a049496"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.193353 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c2g2r" event={"ID":"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6","Type":"ContainerStarted","Data":"efac50fa2c171a3ffac9514f2b79a3f7d9e722213d7a45a4717d5568fbb2cabf"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.193415 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c2g2r" event={"ID":"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6","Type":"ContainerStarted","Data":"3841be5a59d048e0af33a74ae262799efaaf64a36364282614578b569b6d53c2"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.207374 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6558fc797b-2qp4r" event={"ID":"1944527b-6ffc-4b6b-954f-4b01394ea0cf","Type":"ContainerStarted","Data":"8ac4f70641365c40d7d066cd108525fd104f0eaf62202401f21b85c5a5995efc"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.207420 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6558fc797b-2qp4r" event={"ID":"1944527b-6ffc-4b6b-954f-4b01394ea0cf","Type":"ContainerStarted","Data":"cf40a7ca87ed0a4d9665b170715e4da6593ca6539ad9725e2b06d18b9f708490"} Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.300654 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-c2g2r" podStartSLOduration=31.300629754 podStartE2EDuration="31.300629754s" podCreationTimestamp="2025-11-25 10:54:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:30.250319329 +0000 UTC m=+1127.063935916" watchObservedRunningTime="2025-11-25 10:55:30.300629754 +0000 UTC m=+1127.114246341" Nov 25 10:55:30 crc kubenswrapper[4696]: I1125 10:55:30.338463 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6558fc797b-2qp4r" podStartSLOduration=47.82148271 podStartE2EDuration="48.338442456s" podCreationTimestamp="2025-11-25 10:54:42 +0000 UTC" firstStartedPulling="2025-11-25 10:55:28.959319402 +0000 UTC m=+1125.772935989" lastFinishedPulling="2025-11-25 10:55:29.476279148 +0000 UTC m=+1126.289895735" observedRunningTime="2025-11-25 10:55:30.299827728 +0000 UTC m=+1127.113444315" watchObservedRunningTime="2025-11-25 10:55:30.338442456 +0000 UTC m=+1127.152059043" Nov 25 10:55:31 crc kubenswrapper[4696]: I1125 10:55:31.223235 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerStarted","Data":"b743c34bdb2c734c6424f22dad595eb14070e34899fbea0e6b459832b98e3148"} Nov 25 10:55:31 crc kubenswrapper[4696]: I1125 10:55:31.236290 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" event={"ID":"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a","Type":"ContainerStarted","Data":"82ab43d8f8da99faf8b849727008c587b3003c2fd55a87177fed8af175189d52"} Nov 25 10:55:31 crc kubenswrapper[4696]: I1125 10:55:31.236342 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:55:31 crc kubenswrapper[4696]: I1125 10:55:31.253442 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c8fd89c8d-5vb5f" podStartSLOduration=48.644933348 podStartE2EDuration="49.253423528s" podCreationTimestamp="2025-11-25 10:54:42 +0000 UTC" firstStartedPulling="2025-11-25 10:55:28.867410946 +0000 UTC m=+1125.681027533" lastFinishedPulling="2025-11-25 10:55:29.475901126 +0000 UTC m=+1126.289517713" observedRunningTime="2025-11-25 10:55:31.242558518 +0000 UTC m=+1128.056175115" watchObservedRunningTime="2025-11-25 10:55:31.253423528 +0000 UTC m=+1128.067040115" Nov 25 10:55:31 crc kubenswrapper[4696]: I1125 10:55:31.263119 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerStarted","Data":"b9175f8833ce38cbf7619b6d702ee534c161e387b5ab0ec86cc94cf0fc3e71ee"} Nov 25 10:55:31 crc kubenswrapper[4696]: I1125 10:55:31.283430 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" podStartSLOduration=35.283413377 podStartE2EDuration="35.283413377s" podCreationTimestamp="2025-11-25 10:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:31.282193888 +0000 UTC m=+1128.095810485" watchObservedRunningTime="2025-11-25 10:55:31.283413377 +0000 UTC m=+1128.097029964" Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.291484 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc1c6f7a-035e-420f-a056-f63d0c8babcd","Type":"ContainerStarted","Data":"2d760f28e650f4098f8e3229e577204ddc81f100d68080c0b57139eb909f094a"} Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.291890 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-log" containerID="cri-o://3596bf5ef27d2a873057fc6aab0e0028ec618e63f6bf238d827b811342547d38" gracePeriod=30 Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.292054 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-httpd" containerID="cri-o://2d760f28e650f4098f8e3229e577204ddc81f100d68080c0b57139eb909f094a" gracePeriod=30 Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.298243 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9","Type":"ContainerStarted","Data":"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276"} Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.298357 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-log" containerID="cri-o://f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d" gracePeriod=30 Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.298385 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-httpd" containerID="cri-o://155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276" gracePeriod=30 Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.320272 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=36.320252781 podStartE2EDuration="36.320252781s" podCreationTimestamp="2025-11-25 10:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:32.313547612 +0000 UTC m=+1129.127164199" watchObservedRunningTime="2025-11-25 10:55:32.320252781 +0000 UTC m=+1129.133869358" Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.349483 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=35.349464985 podStartE2EDuration="35.349464985s" podCreationTimestamp="2025-11-25 10:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:32.348592548 +0000 UTC m=+1129.162209125" watchObservedRunningTime="2025-11-25 10:55:32.349464985 +0000 UTC m=+1129.163081572" Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.760964 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.761037 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.876955 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:55:32 crc kubenswrapper[4696]: I1125 10:55:32.877233 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.352855 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.362695 4696 generic.go:334] "Generic (PLEG): container finished" podID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" containerID="5c97bb9514113109c93a46209e8213df0786b15a142aff872fbb81cc5256e3a0" exitCode=0 Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.363346 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dq4dk" event={"ID":"c1bfd7e7-4caa-4b79-be66-004376b3f7f4","Type":"ContainerDied","Data":"5c97bb9514113109c93a46209e8213df0786b15a142aff872fbb81cc5256e3a0"} Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.392601 4696 generic.go:334] "Generic (PLEG): container finished" podID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerID="2d760f28e650f4098f8e3229e577204ddc81f100d68080c0b57139eb909f094a" exitCode=0 Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.392641 4696 generic.go:334] "Generic (PLEG): container finished" podID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerID="3596bf5ef27d2a873057fc6aab0e0028ec618e63f6bf238d827b811342547d38" exitCode=143 Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.392730 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc1c6f7a-035e-420f-a056-f63d0c8babcd","Type":"ContainerDied","Data":"2d760f28e650f4098f8e3229e577204ddc81f100d68080c0b57139eb909f094a"} Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.392768 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc1c6f7a-035e-420f-a056-f63d0c8babcd","Type":"ContainerDied","Data":"3596bf5ef27d2a873057fc6aab0e0028ec618e63f6bf238d827b811342547d38"} Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.394618 4696 generic.go:334] "Generic (PLEG): container finished" podID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerID="155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276" exitCode=0 Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.394649 4696 generic.go:334] "Generic (PLEG): container finished" podID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerID="f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d" exitCode=143 Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.396615 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.396781 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9","Type":"ContainerDied","Data":"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276"} Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.396805 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9","Type":"ContainerDied","Data":"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d"} Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.396815 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9","Type":"ContainerDied","Data":"95b71d0f28bde65752d6a6d7bccf95dce466324ed92b11b440984dda4947b6fc"} Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.396830 4696 scope.go:117] "RemoveContainer" containerID="155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.471821 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.471878 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwm2f\" (UniqueName: \"kubernetes.io/projected/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-kube-api-access-bwm2f\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.471921 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-scripts\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.471995 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-httpd-run\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.472054 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-logs\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.472070 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-combined-ca-bundle\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.472098 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-config-data\") pod \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\" (UID: \"093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.478072 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-logs" (OuterVolumeSpecName: "logs") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.478230 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.483443 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.485792 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-scripts" (OuterVolumeSpecName: "scripts") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.495242 4696 scope.go:117] "RemoveContainer" containerID="f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.514835 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-kube-api-access-bwm2f" (OuterVolumeSpecName: "kube-api-access-bwm2f") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "kube-api-access-bwm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.574311 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.574510 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.574577 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.574630 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwm2f\" (UniqueName: \"kubernetes.io/projected/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-kube-api-access-bwm2f\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.574710 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.590125 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.594196 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.633417 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-config-data" (OuterVolumeSpecName: "config-data") pod "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" (UID: "093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.676053 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.676093 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.676104 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.686635 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.762344 4696 scope.go:117] "RemoveContainer" containerID="155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.762858 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276\": container with ID starting with 155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276 not found: ID does not exist" containerID="155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.762897 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276"} err="failed to get container status \"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276\": rpc error: code = NotFound desc = could not find container \"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276\": container with ID starting with 155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276 not found: ID does not exist" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.762930 4696 scope.go:117] "RemoveContainer" containerID="f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.769890 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d\": container with ID starting with f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d not found: ID does not exist" containerID="f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.769946 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d"} err="failed to get container status \"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d\": rpc error: code = NotFound desc = could not find container \"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d\": container with ID starting with f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d not found: ID does not exist" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.769981 4696 scope.go:117] "RemoveContainer" containerID="155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.770852 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276"} err="failed to get container status \"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276\": rpc error: code = NotFound desc = could not find container \"155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276\": container with ID starting with 155dc73dc060dd97ccea353c58b3c6d023481c1de8345c7e30c1349ceb80c276 not found: ID does not exist" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.770883 4696 scope.go:117] "RemoveContainer" containerID="f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.772525 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.776134 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d"} err="failed to get container status \"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d\": rpc error: code = NotFound desc = could not find container \"f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d\": container with ID starting with f8905646db02430ada2e500701e786413de2673af8cc52921a8aec5890367c6d not found: ID does not exist" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.776852 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.776906 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-combined-ca-bundle\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.777026 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-config-data\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.777060 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-scripts\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.777088 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-logs\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.777116 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-httpd-run\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.777147 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqhg6\" (UniqueName: \"kubernetes.io/projected/cc1c6f7a-035e-420f-a056-f63d0c8babcd-kube-api-access-nqhg6\") pod \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\" (UID: \"cc1c6f7a-035e-420f-a056-f63d0c8babcd\") " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.795002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-logs" (OuterVolumeSpecName: "logs") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.795372 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.795570 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc1c6f7a-035e-420f-a056-f63d0c8babcd-kube-api-access-nqhg6" (OuterVolumeSpecName: "kube-api-access-nqhg6") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "kube-api-access-nqhg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.803848 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.805587 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.810703 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-scripts" (OuterVolumeSpecName: "scripts") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.831760 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.832105 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832119 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.832131 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-httpd" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832137 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-httpd" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.832159 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="init" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832164 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="init" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.832174 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-log" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832179 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-log" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.832192 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-httpd" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832197 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-httpd" Nov 25 10:55:33 crc kubenswrapper[4696]: E1125 10:55:33.832214 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-log" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832220 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-log" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832365 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-log" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832374 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-log" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832382 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" containerName="glance-httpd" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832388 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aac8f88-5b55-4f25-8cee-6b20fa9faecd" containerName="dnsmasq-dns" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.832406 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" containerName="glance-httpd" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.833372 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.839375 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.839920 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.848820 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.874684 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.878732 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.878760 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.878768 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc1c6f7a-035e-420f-a056-f63d0c8babcd-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.878778 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqhg6\" (UniqueName: \"kubernetes.io/projected/cc1c6f7a-035e-420f-a056-f63d0c8babcd-kube-api-access-nqhg6\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.878811 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.878820 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.899497 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.907709 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-config-data" (OuterVolumeSpecName: "config-data") pod "cc1c6f7a-035e-420f-a056-f63d0c8babcd" (UID: "cc1c6f7a-035e-420f-a056-f63d0c8babcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980579 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980600 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980648 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980747 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgq6n\" (UniqueName: \"kubernetes.io/projected/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-kube-api-access-kgq6n\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980771 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-logs\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980790 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980840 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:33 crc kubenswrapper[4696]: I1125 10:55:33.980851 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1c6f7a-035e-420f-a056-f63d0c8babcd-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.054987 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9" path="/var/lib/kubelet/pods/093f7cf0-b4bc-440f-8c5f-9bd9949bf4a9/volumes" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082591 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgq6n\" (UniqueName: \"kubernetes.io/projected/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-kube-api-access-kgq6n\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082643 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-logs\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082690 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082799 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082824 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082857 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.082882 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.083125 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-logs\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.083255 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.083434 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.092290 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.094373 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.095392 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.101814 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.104331 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgq6n\" (UniqueName: \"kubernetes.io/projected/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-kube-api-access-kgq6n\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.117609 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.170834 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.410136 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.410136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc1c6f7a-035e-420f-a056-f63d0c8babcd","Type":"ContainerDied","Data":"0eb998a38e8d5cfa4964f82d4355fde07b5ccb94e6d6bd6c1b727f8a59d72894"} Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.411109 4696 scope.go:117] "RemoveContainer" containerID="2d760f28e650f4098f8e3229e577204ddc81f100d68080c0b57139eb909f094a" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.549776 4696 scope.go:117] "RemoveContainer" containerID="3596bf5ef27d2a873057fc6aab0e0028ec618e63f6bf238d827b811342547d38" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.566959 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.604981 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.637727 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.639514 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.646525 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.646583 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.649832 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.711367 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.711741 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc6hg\" (UniqueName: \"kubernetes.io/projected/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-kube-api-access-wc6hg\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.711800 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.711864 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.711928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.711955 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.712086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-logs\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.712115 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.722937 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.821761 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc6hg\" (UniqueName: \"kubernetes.io/projected/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-kube-api-access-wc6hg\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.821816 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.821855 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.821899 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.821925 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.821998 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-logs\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.822079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.822122 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.824048 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.828275 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-logs\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.829862 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.851150 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-config-data\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.861165 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.862061 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-scripts\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.868531 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc6hg\" (UniqueName: \"kubernetes.io/projected/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-kube-api-access-wc6hg\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.870983 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.899417 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " pod="openstack/glance-default-external-api-0" Nov 25 10:55:34 crc kubenswrapper[4696]: I1125 10:55:34.983970 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:55:35 crc kubenswrapper[4696]: I1125 10:55:35.426288 4696 generic.go:334] "Generic (PLEG): container finished" podID="e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" containerID="efac50fa2c171a3ffac9514f2b79a3f7d9e722213d7a45a4717d5568fbb2cabf" exitCode=0 Nov 25 10:55:35 crc kubenswrapper[4696]: I1125 10:55:35.426444 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c2g2r" event={"ID":"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6","Type":"ContainerDied","Data":"efac50fa2c171a3ffac9514f2b79a3f7d9e722213d7a45a4717d5568fbb2cabf"} Nov 25 10:55:36 crc kubenswrapper[4696]: I1125 10:55:36.052623 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc1c6f7a-035e-420f-a056-f63d0c8babcd" path="/var/lib/kubelet/pods/cc1c6f7a-035e-420f-a056-f63d0c8babcd/volumes" Nov 25 10:55:37 crc kubenswrapper[4696]: I1125 10:55:37.325958 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:55:37 crc kubenswrapper[4696]: I1125 10:55:37.399472 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-7d95t"] Nov 25 10:55:37 crc kubenswrapper[4696]: I1125 10:55:37.399774 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" podUID="281499bc-198d-46c6-82e6-ee9260834650" containerName="dnsmasq-dns" containerID="cri-o://83f739bd34ab4041c7712194bde2e25bc9a44af6942bd435438bd50ef7bf6b8e" gracePeriod=10 Nov 25 10:55:38 crc kubenswrapper[4696]: W1125 10:55:38.437843 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14c5ee2e_23a6_411e_9b8f_2cc34bc22daf.slice/crio-b8ddae61627bc53cab526ce188526e19de95b865681eabea42bf9e6e3ce6268f WatchSource:0}: Error finding container b8ddae61627bc53cab526ce188526e19de95b865681eabea42bf9e6e3ce6268f: Status 404 returned error can't find the container with id b8ddae61627bc53cab526ce188526e19de95b865681eabea42bf9e6e3ce6268f Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.484324 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf","Type":"ContainerStarted","Data":"b8ddae61627bc53cab526ce188526e19de95b865681eabea42bf9e6e3ce6268f"} Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.485888 4696 generic.go:334] "Generic (PLEG): container finished" podID="281499bc-198d-46c6-82e6-ee9260834650" containerID="83f739bd34ab4041c7712194bde2e25bc9a44af6942bd435438bd50ef7bf6b8e" exitCode=0 Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.485935 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" event={"ID":"281499bc-198d-46c6-82e6-ee9260834650","Type":"ContainerDied","Data":"83f739bd34ab4041c7712194bde2e25bc9a44af6942bd435438bd50ef7bf6b8e"} Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.487509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c2g2r" event={"ID":"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6","Type":"ContainerDied","Data":"3841be5a59d048e0af33a74ae262799efaaf64a36364282614578b569b6d53c2"} Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.487538 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3841be5a59d048e0af33a74ae262799efaaf64a36364282614578b569b6d53c2" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.492521 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dq4dk" event={"ID":"c1bfd7e7-4caa-4b79-be66-004376b3f7f4","Type":"ContainerDied","Data":"7e6242f99070ea873b9371f40e767753f7fb49abd2bc83a6ea336072fb84dd9d"} Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.492553 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e6242f99070ea873b9371f40e767753f7fb49abd2bc83a6ea336072fb84dd9d" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.525993 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dq4dk" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.530218 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.624369 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-logs\") pod \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.625982 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-scripts\") pod \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626056 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-config-data\") pod \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626082 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5bhh\" (UniqueName: \"kubernetes.io/projected/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-kube-api-access-b5bhh\") pod \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626170 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-credential-keys\") pod \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626203 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-combined-ca-bundle\") pod \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626231 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-scripts\") pod \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626251 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-config-data\") pod \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626272 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-fernet-keys\") pod \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\" (UID: \"e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626310 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-combined-ca-bundle\") pod \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.626349 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2c2c\" (UniqueName: \"kubernetes.io/projected/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-kube-api-access-d2c2c\") pod \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\" (UID: \"c1bfd7e7-4caa-4b79-be66-004376b3f7f4\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.625809 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-logs" (OuterVolumeSpecName: "logs") pod "c1bfd7e7-4caa-4b79-be66-004376b3f7f4" (UID: "c1bfd7e7-4caa-4b79-be66-004376b3f7f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.643572 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" (UID: "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.643762 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-kube-api-access-b5bhh" (OuterVolumeSpecName: "kube-api-access-b5bhh") pod "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" (UID: "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6"). InnerVolumeSpecName "kube-api-access-b5bhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.646542 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-scripts" (OuterVolumeSpecName: "scripts") pod "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" (UID: "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.656050 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-scripts" (OuterVolumeSpecName: "scripts") pod "c1bfd7e7-4caa-4b79-be66-004376b3f7f4" (UID: "c1bfd7e7-4caa-4b79-be66-004376b3f7f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.672966 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-kube-api-access-d2c2c" (OuterVolumeSpecName: "kube-api-access-d2c2c") pod "c1bfd7e7-4caa-4b79-be66-004376b3f7f4" (UID: "c1bfd7e7-4caa-4b79-be66-004376b3f7f4"). InnerVolumeSpecName "kube-api-access-d2c2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.681761 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" (UID: "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.721696 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-config-data" (OuterVolumeSpecName: "config-data") pod "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" (UID: "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.721745 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-config-data" (OuterVolumeSpecName: "config-data") pod "c1bfd7e7-4caa-4b79-be66-004376b3f7f4" (UID: "c1bfd7e7-4caa-4b79-be66-004376b3f7f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728491 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728521 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728534 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5bhh\" (UniqueName: \"kubernetes.io/projected/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-kube-api-access-b5bhh\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728545 4696 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728553 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728562 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728569 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728580 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2c2c\" (UniqueName: \"kubernetes.io/projected/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-kube-api-access-d2c2c\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.728587 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.729608 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1bfd7e7-4caa-4b79-be66-004376b3f7f4" (UID: "c1bfd7e7-4caa-4b79-be66-004376b3f7f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.734888 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" (UID: "e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.761712 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.831182 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.831220 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1bfd7e7-4caa-4b79-be66-004376b3f7f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.932219 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-config\") pod \"281499bc-198d-46c6-82e6-ee9260834650\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.932498 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-nb\") pod \"281499bc-198d-46c6-82e6-ee9260834650\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.932534 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-swift-storage-0\") pod \"281499bc-198d-46c6-82e6-ee9260834650\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.932758 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-sb\") pod \"281499bc-198d-46c6-82e6-ee9260834650\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.932780 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-svc\") pod \"281499bc-198d-46c6-82e6-ee9260834650\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.932842 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zp6m\" (UniqueName: \"kubernetes.io/projected/281499bc-198d-46c6-82e6-ee9260834650-kube-api-access-9zp6m\") pod \"281499bc-198d-46c6-82e6-ee9260834650\" (UID: \"281499bc-198d-46c6-82e6-ee9260834650\") " Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.939636 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281499bc-198d-46c6-82e6-ee9260834650-kube-api-access-9zp6m" (OuterVolumeSpecName: "kube-api-access-9zp6m") pod "281499bc-198d-46c6-82e6-ee9260834650" (UID: "281499bc-198d-46c6-82e6-ee9260834650"). InnerVolumeSpecName "kube-api-access-9zp6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.996106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-config" (OuterVolumeSpecName: "config") pod "281499bc-198d-46c6-82e6-ee9260834650" (UID: "281499bc-198d-46c6-82e6-ee9260834650"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:38 crc kubenswrapper[4696]: I1125 10:55:38.999714 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "281499bc-198d-46c6-82e6-ee9260834650" (UID: "281499bc-198d-46c6-82e6-ee9260834650"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.014651 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "281499bc-198d-46c6-82e6-ee9260834650" (UID: "281499bc-198d-46c6-82e6-ee9260834650"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.024577 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "281499bc-198d-46c6-82e6-ee9260834650" (UID: "281499bc-198d-46c6-82e6-ee9260834650"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.035019 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.035043 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.035052 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zp6m\" (UniqueName: \"kubernetes.io/projected/281499bc-198d-46c6-82e6-ee9260834650-kube-api-access-9zp6m\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.035063 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.035071 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.049431 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "281499bc-198d-46c6-82e6-ee9260834650" (UID: "281499bc-198d-46c6-82e6-ee9260834650"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.140950 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/281499bc-198d-46c6-82e6-ee9260834650-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.365915 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:55:39 crc kubenswrapper[4696]: W1125 10:55:39.381399 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac5f4fea_e9ee_4fa9_b919_ba3a1bf21b42.slice/crio-23d176541e9c979c5e5c57cacc68d845ce30274a64abc9327dee79b78bd75930 WatchSource:0}: Error finding container 23d176541e9c979c5e5c57cacc68d845ce30274a64abc9327dee79b78bd75930: Status 404 returned error can't find the container with id 23d176541e9c979c5e5c57cacc68d845ce30274a64abc9327dee79b78bd75930 Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.531312 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" event={"ID":"281499bc-198d-46c6-82e6-ee9260834650","Type":"ContainerDied","Data":"2fb85b74e0b89ab63453ccadbe0efb8532980e1b2359c79c0c09b81777853742"} Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.531356 4696 scope.go:117] "RemoveContainer" containerID="83f739bd34ab4041c7712194bde2e25bc9a44af6942bd435438bd50ef7bf6b8e" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.531471 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-7d95t" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.543306 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerStarted","Data":"b2cff43ffe356daf4428055958f7db69c82732c4acd68ddc2c6dafad540cceb5"} Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.574889 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42","Type":"ContainerStarted","Data":"23d176541e9c979c5e5c57cacc68d845ce30274a64abc9327dee79b78bd75930"} Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.582446 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-7d95t"] Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.588747 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vqmj" event={"ID":"cacd4fa8-b70a-451d-92a9-40e53ea0a406","Type":"ContainerStarted","Data":"00f8873d95d129a119835fee128c27e00b740b6dcaadce41e01b6bdcd3f20925"} Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.583539 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dq4dk" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.583607 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c2g2r" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.635762 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-7d95t"] Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.652191 4696 scope.go:117] "RemoveContainer" containerID="7f393b0fbfc79962968f4bf19ff963227a59be385892bae92cbb1d30a4b13350" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.699179 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5vqmj" podStartSLOduration=3.583602139 podStartE2EDuration="1m7.699155621s" podCreationTimestamp="2025-11-25 10:54:32 +0000 UTC" firstStartedPulling="2025-11-25 10:54:34.839826524 +0000 UTC m=+1071.653443111" lastFinishedPulling="2025-11-25 10:55:38.955380016 +0000 UTC m=+1135.768996593" observedRunningTime="2025-11-25 10:55:39.652074997 +0000 UTC m=+1136.465691584" watchObservedRunningTime="2025-11-25 10:55:39.699155621 +0000 UTC m=+1136.512772208" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.722901 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-646fccfb54-pbhr5"] Nov 25 10:55:39 crc kubenswrapper[4696]: E1125 10:55:39.723285 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" containerName="placement-db-sync" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723311 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" containerName="placement-db-sync" Nov 25 10:55:39 crc kubenswrapper[4696]: E1125 10:55:39.723330 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281499bc-198d-46c6-82e6-ee9260834650" containerName="init" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723336 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="281499bc-198d-46c6-82e6-ee9260834650" containerName="init" Nov 25 10:55:39 crc kubenswrapper[4696]: E1125 10:55:39.723361 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" containerName="keystone-bootstrap" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723367 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" containerName="keystone-bootstrap" Nov 25 10:55:39 crc kubenswrapper[4696]: E1125 10:55:39.723383 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281499bc-198d-46c6-82e6-ee9260834650" containerName="dnsmasq-dns" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723389 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="281499bc-198d-46c6-82e6-ee9260834650" containerName="dnsmasq-dns" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723540 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="281499bc-198d-46c6-82e6-ee9260834650" containerName="dnsmasq-dns" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723554 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" containerName="keystone-bootstrap" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.723576 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" containerName="placement-db-sync" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.724827 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.727343 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-445jl" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.734038 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.734427 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.734707 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.736630 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.764033 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-646fccfb54-pbhr5"] Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.791458 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5547bcb8fc-b65gm"] Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.792943 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.798615 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.798817 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.798880 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.799403 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r7vqt" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.799703 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.799727 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.825479 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5547bcb8fc-b65gm"] Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858567 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-scripts\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858645 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-combined-ca-bundle\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858850 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-internal-tls-certs\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858890 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-credential-keys\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858919 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-config-data\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858948 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-combined-ca-bundle\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.858975 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-scripts\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859035 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8snvn\" (UniqueName: \"kubernetes.io/projected/537410db-3589-4c1a-bb28-1f0a66a20ff6-kube-api-access-8snvn\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859078 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537410db-3589-4c1a-bb28-1f0a66a20ff6-logs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859116 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-public-tls-certs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859145 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-fernet-keys\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859184 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhqvz\" (UniqueName: \"kubernetes.io/projected/ad61fdbd-b908-4430-8b43-e08260f0dbd1-kube-api-access-rhqvz\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859279 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-config-data\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859307 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-public-tls-certs\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.859365 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-internal-tls-certs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960696 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8snvn\" (UniqueName: \"kubernetes.io/projected/537410db-3589-4c1a-bb28-1f0a66a20ff6-kube-api-access-8snvn\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960745 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537410db-3589-4c1a-bb28-1f0a66a20ff6-logs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960775 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-public-tls-certs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960801 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-fernet-keys\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960831 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhqvz\" (UniqueName: \"kubernetes.io/projected/ad61fdbd-b908-4430-8b43-e08260f0dbd1-kube-api-access-rhqvz\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960853 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-config-data\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960869 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-public-tls-certs\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960893 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-internal-tls-certs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.960918 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-scripts\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.963237 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-combined-ca-bundle\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.963334 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-internal-tls-certs\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.963374 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-credential-keys\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.963401 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-config-data\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.963433 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-combined-ca-bundle\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.963941 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-scripts\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.964626 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-public-tls-certs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.971003 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-combined-ca-bundle\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.961257 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537410db-3589-4c1a-bb28-1f0a66a20ff6-logs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.971998 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-fernet-keys\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.973260 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-scripts\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.977259 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-credential-keys\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.978060 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-config-data\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.978463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-internal-tls-certs\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.978502 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-internal-tls-certs\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.978748 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-public-tls-certs\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.979190 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-scripts\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.979863 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537410db-3589-4c1a-bb28-1f0a66a20ff6-combined-ca-bundle\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.980576 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad61fdbd-b908-4430-8b43-e08260f0dbd1-config-data\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.983179 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8snvn\" (UniqueName: \"kubernetes.io/projected/537410db-3589-4c1a-bb28-1f0a66a20ff6-kube-api-access-8snvn\") pod \"placement-646fccfb54-pbhr5\" (UID: \"537410db-3589-4c1a-bb28-1f0a66a20ff6\") " pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:39 crc kubenswrapper[4696]: I1125 10:55:39.994160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhqvz\" (UniqueName: \"kubernetes.io/projected/ad61fdbd-b908-4430-8b43-e08260f0dbd1-kube-api-access-rhqvz\") pod \"keystone-5547bcb8fc-b65gm\" (UID: \"ad61fdbd-b908-4430-8b43-e08260f0dbd1\") " pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:40 crc kubenswrapper[4696]: I1125 10:55:40.072071 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="281499bc-198d-46c6-82e6-ee9260834650" path="/var/lib/kubelet/pods/281499bc-198d-46c6-82e6-ee9260834650/volumes" Nov 25 10:55:40 crc kubenswrapper[4696]: I1125 10:55:40.085450 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:40 crc kubenswrapper[4696]: I1125 10:55:40.118321 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:40 crc kubenswrapper[4696]: I1125 10:55:40.606102 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf","Type":"ContainerStarted","Data":"541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b"} Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.002315 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-646fccfb54-pbhr5"] Nov 25 10:55:41 crc kubenswrapper[4696]: W1125 10:55:41.078156 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod537410db_3589_4c1a_bb28_1f0a66a20ff6.slice/crio-e8684eb55a415c23e605e3b97d169be62cb742c8c88801a3f7edaabcaa920428 WatchSource:0}: Error finding container e8684eb55a415c23e605e3b97d169be62cb742c8c88801a3f7edaabcaa920428: Status 404 returned error can't find the container with id e8684eb55a415c23e605e3b97d169be62cb742c8c88801a3f7edaabcaa920428 Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.178634 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5547bcb8fc-b65gm"] Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.642777 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf","Type":"ContainerStarted","Data":"84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae"} Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.660894 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42","Type":"ContainerStarted","Data":"5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40"} Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.676195 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.676174284 podStartE2EDuration="8.676174284s" podCreationTimestamp="2025-11-25 10:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:41.662531568 +0000 UTC m=+1138.476148155" watchObservedRunningTime="2025-11-25 10:55:41.676174284 +0000 UTC m=+1138.489790871" Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.693931 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-646fccfb54-pbhr5" event={"ID":"537410db-3589-4c1a-bb28-1f0a66a20ff6","Type":"ContainerStarted","Data":"a9cf8fc9f221a3e07b1b56252bf52b273459f408017a0acfc9f3addde88e2c7b"} Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.693974 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-646fccfb54-pbhr5" event={"ID":"537410db-3589-4c1a-bb28-1f0a66a20ff6","Type":"ContainerStarted","Data":"e8684eb55a415c23e605e3b97d169be62cb742c8c88801a3f7edaabcaa920428"} Nov 25 10:55:41 crc kubenswrapper[4696]: I1125 10:55:41.706512 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5547bcb8fc-b65gm" event={"ID":"ad61fdbd-b908-4430-8b43-e08260f0dbd1","Type":"ContainerStarted","Data":"0edfe2db9014a539d4f7f86f0d670f839fbde7856317ea83f5ccf34945e8c139"} Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.718437 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-646fccfb54-pbhr5" event={"ID":"537410db-3589-4c1a-bb28-1f0a66a20ff6","Type":"ContainerStarted","Data":"8ad0a717b8317413124c57757b96715c05dcaa481ddfb15f8f03d79d50b345b7"} Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.718898 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.722654 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7qtsq" event={"ID":"9353d93a-d886-43d0-a4d4-6ad17a28f52a","Type":"ContainerStarted","Data":"0e534c69c8ef2d7aa45777809c98dc906cf2231b90e8f435f7592bb0301cef72"} Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.731054 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5547bcb8fc-b65gm" event={"ID":"ad61fdbd-b908-4430-8b43-e08260f0dbd1","Type":"ContainerStarted","Data":"fe43ffd40dec3dcf6cd74459e32adabb4fc2a418f4234c6d0f6858e05c488f34"} Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.731185 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.736569 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42","Type":"ContainerStarted","Data":"dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff"} Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.750110 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-646fccfb54-pbhr5" podStartSLOduration=3.7500936400000002 podStartE2EDuration="3.75009364s" podCreationTimestamp="2025-11-25 10:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:42.740372245 +0000 UTC m=+1139.553988832" watchObservedRunningTime="2025-11-25 10:55:42.75009364 +0000 UTC m=+1139.563710227" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.766826 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.780802 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5547bcb8fc-b65gm" podStartSLOduration=3.78078396 podStartE2EDuration="3.78078396s" podCreationTimestamp="2025-11-25 10:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:42.767105342 +0000 UTC m=+1139.580721929" watchObservedRunningTime="2025-11-25 10:55:42.78078396 +0000 UTC m=+1139.594400547" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.817242 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.81721994 podStartE2EDuration="8.81721994s" podCreationTimestamp="2025-11-25 10:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:42.797066929 +0000 UTC m=+1139.610683516" watchObservedRunningTime="2025-11-25 10:55:42.81721994 +0000 UTC m=+1139.630836527" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.844604 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7qtsq" podStartSLOduration=4.020817011 podStartE2EDuration="1m9.844585816s" podCreationTimestamp="2025-11-25 10:54:33 +0000 UTC" firstStartedPulling="2025-11-25 10:54:34.893309048 +0000 UTC m=+1071.706925635" lastFinishedPulling="2025-11-25 10:55:40.717077853 +0000 UTC m=+1137.530694440" observedRunningTime="2025-11-25 10:55:42.824110155 +0000 UTC m=+1139.637726762" watchObservedRunningTime="2025-11-25 10:55:42.844585816 +0000 UTC m=+1139.658202403" Nov 25 10:55:42 crc kubenswrapper[4696]: I1125 10:55:42.878501 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6558fc797b-2qp4r" podUID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 10:55:43 crc kubenswrapper[4696]: I1125 10:55:43.745621 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.171845 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.172255 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.212710 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.224510 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.752809 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.753122 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.985346 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:55:44 crc kubenswrapper[4696]: I1125 10:55:44.985389 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:55:45 crc kubenswrapper[4696]: I1125 10:55:45.028494 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:55:45 crc kubenswrapper[4696]: I1125 10:55:45.093543 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:55:45 crc kubenswrapper[4696]: I1125 10:55:45.761414 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:55:45 crc kubenswrapper[4696]: I1125 10:55:45.761723 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:55:46 crc kubenswrapper[4696]: I1125 10:55:46.770189 4696 generic.go:334] "Generic (PLEG): container finished" podID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" containerID="00f8873d95d129a119835fee128c27e00b740b6dcaadce41e01b6bdcd3f20925" exitCode=0 Nov 25 10:55:46 crc kubenswrapper[4696]: I1125 10:55:46.770271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vqmj" event={"ID":"cacd4fa8-b70a-451d-92a9-40e53ea0a406","Type":"ContainerDied","Data":"00f8873d95d129a119835fee128c27e00b740b6dcaadce41e01b6bdcd3f20925"} Nov 25 10:55:47 crc kubenswrapper[4696]: I1125 10:55:47.797440 4696 generic.go:334] "Generic (PLEG): container finished" podID="3a052a79-e780-430c-a049-cc01418ba194" containerID="aea37d144e9450979ed996b37ced91a9f9660ed6c748e718a17f0296d80bb5ad" exitCode=0 Nov 25 10:55:47 crc kubenswrapper[4696]: I1125 10:55:47.797534 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqj4p" event={"ID":"3a052a79-e780-430c-a049-cc01418ba194","Type":"ContainerDied","Data":"aea37d144e9450979ed996b37ced91a9f9660ed6c748e718a17f0296d80bb5ad"} Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.845328 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqj4p" event={"ID":"3a052a79-e780-430c-a049-cc01418ba194","Type":"ContainerDied","Data":"9217b49b3b10ae613e787924163339a7611648d6e4863b20b1f5234a50ed417f"} Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.845524 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9217b49b3b10ae613e787924163339a7611648d6e4863b20b1f5234a50ed417f" Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.859621 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vqmj" event={"ID":"cacd4fa8-b70a-451d-92a9-40e53ea0a406","Type":"ContainerDied","Data":"b3cb50826db21c55197adaf101932c8361793a73d6ea060bd00bac195a66b9d3"} Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.859671 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3cb50826db21c55197adaf101932c8361793a73d6ea060bd00bac195a66b9d3" Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.900732 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.941850 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.958729 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5xg9\" (UniqueName: \"kubernetes.io/projected/cacd4fa8-b70a-451d-92a9-40e53ea0a406-kube-api-access-k5xg9\") pod \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.958824 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-combined-ca-bundle\") pod \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.958970 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-db-sync-config-data\") pod \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\" (UID: \"cacd4fa8-b70a-451d-92a9-40e53ea0a406\") " Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.965808 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cacd4fa8-b70a-451d-92a9-40e53ea0a406-kube-api-access-k5xg9" (OuterVolumeSpecName: "kube-api-access-k5xg9") pod "cacd4fa8-b70a-451d-92a9-40e53ea0a406" (UID: "cacd4fa8-b70a-451d-92a9-40e53ea0a406"). InnerVolumeSpecName "kube-api-access-k5xg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.974205 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cacd4fa8-b70a-451d-92a9-40e53ea0a406" (UID: "cacd4fa8-b70a-451d-92a9-40e53ea0a406"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:50 crc kubenswrapper[4696]: I1125 10:55:50.995087 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cacd4fa8-b70a-451d-92a9-40e53ea0a406" (UID: "cacd4fa8-b70a-451d-92a9-40e53ea0a406"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.061866 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-config\") pod \"3a052a79-e780-430c-a049-cc01418ba194\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.061933 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-combined-ca-bundle\") pod \"3a052a79-e780-430c-a049-cc01418ba194\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.062166 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r86gc\" (UniqueName: \"kubernetes.io/projected/3a052a79-e780-430c-a049-cc01418ba194-kube-api-access-r86gc\") pod \"3a052a79-e780-430c-a049-cc01418ba194\" (UID: \"3a052a79-e780-430c-a049-cc01418ba194\") " Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.063452 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5xg9\" (UniqueName: \"kubernetes.io/projected/cacd4fa8-b70a-451d-92a9-40e53ea0a406-kube-api-access-k5xg9\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.063542 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.063855 4696 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cacd4fa8-b70a-451d-92a9-40e53ea0a406-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.073369 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a052a79-e780-430c-a049-cc01418ba194-kube-api-access-r86gc" (OuterVolumeSpecName: "kube-api-access-r86gc") pod "3a052a79-e780-430c-a049-cc01418ba194" (UID: "3a052a79-e780-430c-a049-cc01418ba194"). InnerVolumeSpecName "kube-api-access-r86gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.096819 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a052a79-e780-430c-a049-cc01418ba194" (UID: "3a052a79-e780-430c-a049-cc01418ba194"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.116890 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-config" (OuterVolumeSpecName: "config") pod "3a052a79-e780-430c-a049-cc01418ba194" (UID: "3a052a79-e780-430c-a049-cc01418ba194"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.165131 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r86gc\" (UniqueName: \"kubernetes.io/projected/3a052a79-e780-430c-a049-cc01418ba194-kube-api-access-r86gc\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.165468 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.165482 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a052a79-e780-430c-a049-cc01418ba194-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.785508 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.837650 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.870377 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vqmj" Nov 25 10:55:51 crc kubenswrapper[4696]: I1125 10:55:51.871486 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqj4p" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.365378 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b978c4f56-2cdxx"] Nov 25 10:55:52 crc kubenswrapper[4696]: E1125 10:55:52.366208 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a052a79-e780-430c-a049-cc01418ba194" containerName="neutron-db-sync" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.366226 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a052a79-e780-430c-a049-cc01418ba194" containerName="neutron-db-sync" Nov 25 10:55:52 crc kubenswrapper[4696]: E1125 10:55:52.366235 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" containerName="barbican-db-sync" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.366241 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" containerName="barbican-db-sync" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.366398 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a052a79-e780-430c-a049-cc01418ba194" containerName="neutron-db-sync" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.366418 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" containerName="barbican-db-sync" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.367544 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.382551 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.383398 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-gmkwd" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.385130 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.385613 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-httpd-config\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.385687 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-ovndb-tls-certs\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.385710 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-combined-ca-bundle\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.385728 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-config\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.385742 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zz6m\" (UniqueName: \"kubernetes.io/projected/88ab5f30-bc12-49da-8078-32be5ae2d5d9-kube-api-access-7zz6m\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.388358 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.412081 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-nkk2q"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.413582 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.429616 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b978c4f56-2cdxx"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.440494 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-nkk2q"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.486519 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8dcv\" (UniqueName: \"kubernetes.io/projected/c53b29a3-e64e-459c-a1b8-d700ee54486b-kube-api-access-r8dcv\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.486599 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-ovndb-tls-certs\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.486621 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-combined-ca-bundle\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.486642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-config\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.486669 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zz6m\" (UniqueName: \"kubernetes.io/projected/88ab5f30-bc12-49da-8078-32be5ae2d5d9-kube-api-access-7zz6m\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.487464 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.487516 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-svc\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.487540 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-config\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.487555 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.487603 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-httpd-config\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.487629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.495959 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-config\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.499228 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-combined-ca-bundle\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.499988 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-ovndb-tls-certs\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.513848 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zz6m\" (UniqueName: \"kubernetes.io/projected/88ab5f30-bc12-49da-8078-32be5ae2d5d9-kube-api-access-7zz6m\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.544300 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-httpd-config\") pod \"neutron-b978c4f56-2cdxx\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.588600 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8dcv\" (UniqueName: \"kubernetes.io/projected/c53b29a3-e64e-459c-a1b8-d700ee54486b-kube-api-access-r8dcv\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.588715 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.588764 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-svc\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.588792 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-config\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.588813 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.588868 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.589769 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.589941 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-svc\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.590284 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-config\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.590907 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.590964 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.663948 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8dcv\" (UniqueName: \"kubernetes.io/projected/c53b29a3-e64e-459c-a1b8-d700ee54486b-kube-api-access-r8dcv\") pod \"dnsmasq-dns-6b7b667979-nkk2q\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.677180 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-66cc7b58db-xph79"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.678598 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.684059 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.684274 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.696622 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-64b77754f5-27gpf"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.698279 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.709599 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.711259 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-26hbb" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.725918 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66cc7b58db-xph79"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.730066 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.755964 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.762903 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.792769 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64b77754f5-27gpf"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.794890 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-config-data-custom\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.794933 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k85rz\" (UniqueName: \"kubernetes.io/projected/9e0f5310-cf60-4125-9b91-2dfda5383c7d-kube-api-access-k85rz\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.794981 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-combined-ca-bundle\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.795070 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-config-data\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.795105 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0f5310-cf60-4125-9b91-2dfda5383c7d-logs\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.831638 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-nkk2q"] Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.888204 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6558fc797b-2qp4r" podUID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.911286 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-config-data\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.934773 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-config-data\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.940864 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0f5310-cf60-4125-9b91-2dfda5383c7d-logs\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941022 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-combined-ca-bundle\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941082 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-config-data-custom\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941101 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k85rz\" (UniqueName: \"kubernetes.io/projected/9e0f5310-cf60-4125-9b91-2dfda5383c7d-kube-api-access-k85rz\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941163 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c86z9\" (UniqueName: \"kubernetes.io/projected/22dbae09-113b-4136-b9d6-6162815b5ef7-kube-api-access-c86z9\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941184 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-combined-ca-bundle\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941244 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-config-data\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941275 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-config-data-custom\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.941332 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22dbae09-113b-4136-b9d6-6162815b5ef7-logs\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.945197 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0f5310-cf60-4125-9b91-2dfda5383c7d-logs\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.958687 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-config-data-custom\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:52 crc kubenswrapper[4696]: I1125 10:55:52.959532 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e0f5310-cf60-4125-9b91-2dfda5383c7d-combined-ca-bundle\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.006657 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-89b7b674-c7w92"] Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.008183 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.016994 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.024854 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerStarted","Data":"9ed29f07b52672503f9cf758e7c8520c54dd65d393161bd977d1e03aeff4522f"} Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.025031 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-central-agent" containerID="cri-o://b129008fc63a532aaeecc84712dd6d46fb99265b4c627542ad08d8514e68301d" gracePeriod=30 Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.025299 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.025338 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="proxy-httpd" containerID="cri-o://9ed29f07b52672503f9cf758e7c8520c54dd65d393161bd977d1e03aeff4522f" gracePeriod=30 Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.025375 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="sg-core" containerID="cri-o://b2cff43ffe356daf4428055958f7db69c82732c4acd68ddc2c6dafad540cceb5" gracePeriod=30 Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.025408 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-notification-agent" containerID="cri-o://b9175f8833ce38cbf7619b6d702ee534c161e387b5ab0ec86cc94cf0fc3e71ee" gracePeriod=30 Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.030083 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-89b7b674-c7w92"] Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.043463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k85rz\" (UniqueName: \"kubernetes.io/projected/9e0f5310-cf60-4125-9b91-2dfda5383c7d-kube-api-access-k85rz\") pod \"barbican-keystone-listener-66cc7b58db-xph79\" (UID: \"9e0f5310-cf60-4125-9b91-2dfda5383c7d\") " pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.061392 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-config-data\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.061442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-config-data-custom\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.061486 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22dbae09-113b-4136-b9d6-6162815b5ef7-logs\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.061569 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-combined-ca-bundle\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.061617 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c86z9\" (UniqueName: \"kubernetes.io/projected/22dbae09-113b-4136-b9d6-6162815b5ef7-kube-api-access-c86z9\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.063455 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22dbae09-113b-4136-b9d6-6162815b5ef7-logs\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.064311 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.073694 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-z7stz"] Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.085852 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.104325 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c86z9\" (UniqueName: \"kubernetes.io/projected/22dbae09-113b-4136-b9d6-6162815b5ef7-kube-api-access-c86z9\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.105124 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-combined-ca-bundle\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.106263 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-config-data\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.154329 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22dbae09-113b-4136-b9d6-6162815b5ef7-config-data-custom\") pod \"barbican-worker-64b77754f5-27gpf\" (UID: \"22dbae09-113b-4136-b9d6-6162815b5ef7\") " pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.207741 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-z7stz"] Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.239870 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b6a593d-529f-4d37-a4cb-c86dd239379e-logs\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240100 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240169 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240260 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data-custom\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240300 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-config\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240357 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240456 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-combined-ca-bundle\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240477 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240512 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmbr8\" (UniqueName: \"kubernetes.io/projected/4b6a593d-529f-4d37-a4cb-c86dd239379e-kube-api-access-lmbr8\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240548 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc9xr\" (UniqueName: \"kubernetes.io/projected/5d51458e-8948-4b70-91ba-02f5a4a323f0-kube-api-access-zc9xr\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.240974 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64b77754f5-27gpf" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.329708 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.511171605 podStartE2EDuration="1m21.32956561s" podCreationTimestamp="2025-11-25 10:54:32 +0000 UTC" firstStartedPulling="2025-11-25 10:54:34.83837556 +0000 UTC m=+1071.651992147" lastFinishedPulling="2025-11-25 10:55:51.656769565 +0000 UTC m=+1148.470386152" observedRunningTime="2025-11-25 10:55:53.185265514 +0000 UTC m=+1149.998882101" watchObservedRunningTime="2025-11-25 10:55:53.32956561 +0000 UTC m=+1150.143182197" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355039 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355109 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355160 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data-custom\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355190 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-config\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355271 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-combined-ca-bundle\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355290 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355320 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmbr8\" (UniqueName: \"kubernetes.io/projected/4b6a593d-529f-4d37-a4cb-c86dd239379e-kube-api-access-lmbr8\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355350 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc9xr\" (UniqueName: \"kubernetes.io/projected/5d51458e-8948-4b70-91ba-02f5a4a323f0-kube-api-access-zc9xr\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355366 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355429 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b6a593d-529f-4d37-a4cb-c86dd239379e-logs\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.355899 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b6a593d-529f-4d37-a4cb-c86dd239379e-logs\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.356584 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.357139 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.367221 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.367908 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-config\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.383241 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.388526 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.389094 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-combined-ca-bundle\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.400318 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data-custom\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.407316 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmbr8\" (UniqueName: \"kubernetes.io/projected/4b6a593d-529f-4d37-a4cb-c86dd239379e-kube-api-access-lmbr8\") pod \"barbican-api-89b7b674-c7w92\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.411368 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc9xr\" (UniqueName: \"kubernetes.io/projected/5d51458e-8948-4b70-91ba-02f5a4a323f0-kube-api-access-zc9xr\") pod \"dnsmasq-dns-848cf88cfc-z7stz\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.497194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.694338 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-nkk2q"] Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.703292 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.841787 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66cc7b58db-xph79"] Nov 25 10:55:53 crc kubenswrapper[4696]: I1125 10:55:53.855958 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b978c4f56-2cdxx"] Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.148215 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" event={"ID":"c53b29a3-e64e-459c-a1b8-d700ee54486b","Type":"ContainerStarted","Data":"7dca901078cb99be9b4f9941f8df086402b7a056b3d868fcaa01fddf05d6c1f5"} Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.194752 4696 generic.go:334] "Generic (PLEG): container finished" podID="96044f36-992b-44ce-89b4-a059b1efc117" containerID="9ed29f07b52672503f9cf758e7c8520c54dd65d393161bd977d1e03aeff4522f" exitCode=0 Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.194783 4696 generic.go:334] "Generic (PLEG): container finished" podID="96044f36-992b-44ce-89b4-a059b1efc117" containerID="b2cff43ffe356daf4428055958f7db69c82732c4acd68ddc2c6dafad540cceb5" exitCode=2 Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.194791 4696 generic.go:334] "Generic (PLEG): container finished" podID="96044f36-992b-44ce-89b4-a059b1efc117" containerID="b129008fc63a532aaeecc84712dd6d46fb99265b4c627542ad08d8514e68301d" exitCode=0 Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.194830 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerDied","Data":"9ed29f07b52672503f9cf758e7c8520c54dd65d393161bd977d1e03aeff4522f"} Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.194862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerDied","Data":"b2cff43ffe356daf4428055958f7db69c82732c4acd68ddc2c6dafad540cceb5"} Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.194872 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerDied","Data":"b129008fc63a532aaeecc84712dd6d46fb99265b4c627542ad08d8514e68301d"} Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.195832 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b978c4f56-2cdxx" event={"ID":"88ab5f30-bc12-49da-8078-32be5ae2d5d9","Type":"ContainerStarted","Data":"8a6a4fe123b1c778b9a903312ab2257947268ebcdee24714d1c006540ffb12c8"} Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.198548 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" event={"ID":"9e0f5310-cf60-4125-9b91-2dfda5383c7d","Type":"ContainerStarted","Data":"dda351fab9721e4d54e7f632545d1321f852df33b1dfd658571a00286144c5d2"} Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.280561 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64b77754f5-27gpf"] Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.568574 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-z7stz"] Nov 25 10:55:54 crc kubenswrapper[4696]: W1125 10:55:54.654170 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d51458e_8948_4b70_91ba_02f5a4a323f0.slice/crio-e8cc8b99d8e21e29dc2081d22cbe189e9a3c2eda2d103713d82a2dbbf5441b54 WatchSource:0}: Error finding container e8cc8b99d8e21e29dc2081d22cbe189e9a3c2eda2d103713d82a2dbbf5441b54: Status 404 returned error can't find the container with id e8cc8b99d8e21e29dc2081d22cbe189e9a3c2eda2d103713d82a2dbbf5441b54 Nov 25 10:55:54 crc kubenswrapper[4696]: I1125 10:55:54.732184 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-89b7b674-c7w92"] Nov 25 10:55:54 crc kubenswrapper[4696]: W1125 10:55:54.742157 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b6a593d_529f_4d37_a4cb_c86dd239379e.slice/crio-f17836c73f8efa8bf62315e96d8965ffd5e297ca65fb7ba4cd3b05afe9272d6e WatchSource:0}: Error finding container f17836c73f8efa8bf62315e96d8965ffd5e297ca65fb7ba4cd3b05afe9272d6e: Status 404 returned error can't find the container with id f17836c73f8efa8bf62315e96d8965ffd5e297ca65fb7ba4cd3b05afe9272d6e Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.242767 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64b77754f5-27gpf" event={"ID":"22dbae09-113b-4136-b9d6-6162815b5ef7","Type":"ContainerStarted","Data":"3ce12921b028e08a1d658b249bdf50cd20891f743199842bff23be49d9ec1675"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.251994 4696 generic.go:334] "Generic (PLEG): container finished" podID="c53b29a3-e64e-459c-a1b8-d700ee54486b" containerID="23501198fceb73df2acc2e09d00dcb299d9596787eb72b296d09c3149386c152" exitCode=0 Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.252069 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" event={"ID":"c53b29a3-e64e-459c-a1b8-d700ee54486b","Type":"ContainerDied","Data":"23501198fceb73df2acc2e09d00dcb299d9596787eb72b296d09c3149386c152"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.267520 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b978c4f56-2cdxx" event={"ID":"88ab5f30-bc12-49da-8078-32be5ae2d5d9","Type":"ContainerStarted","Data":"87d42e215c695a26e3b7b4ca984c8c2baee37d35456521565a5e09c21e03f574"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.267569 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b978c4f56-2cdxx" event={"ID":"88ab5f30-bc12-49da-8078-32be5ae2d5d9","Type":"ContainerStarted","Data":"88a9999603fdc26e8442e67101cb67c0c71a56b11ac4283bd258f1aaf40ea1d6"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.268479 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.274265 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-89b7b674-c7w92" event={"ID":"4b6a593d-529f-4d37-a4cb-c86dd239379e","Type":"ContainerStarted","Data":"d1046fa1331aef9c0bf81217545c4717672a0b12a31fa57d7362160831fb9366"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.274297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-89b7b674-c7w92" event={"ID":"4b6a593d-529f-4d37-a4cb-c86dd239379e","Type":"ContainerStarted","Data":"f17836c73f8efa8bf62315e96d8965ffd5e297ca65fb7ba4cd3b05afe9272d6e"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.293949 4696 generic.go:334] "Generic (PLEG): container finished" podID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerID="40c05b71aa3b47f585a3d1d2dbe93e36b4f218b2c4967de6786b21b251f9941e" exitCode=0 Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.294028 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" event={"ID":"5d51458e-8948-4b70-91ba-02f5a4a323f0","Type":"ContainerDied","Data":"40c05b71aa3b47f585a3d1d2dbe93e36b4f218b2c4967de6786b21b251f9941e"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.294054 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" event={"ID":"5d51458e-8948-4b70-91ba-02f5a4a323f0","Type":"ContainerStarted","Data":"e8cc8b99d8e21e29dc2081d22cbe189e9a3c2eda2d103713d82a2dbbf5441b54"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.297976 4696 generic.go:334] "Generic (PLEG): container finished" podID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" containerID="0e534c69c8ef2d7aa45777809c98dc906cf2231b90e8f435f7592bb0301cef72" exitCode=0 Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.298017 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7qtsq" event={"ID":"9353d93a-d886-43d0-a4d4-6ad17a28f52a","Type":"ContainerDied","Data":"0e534c69c8ef2d7aa45777809c98dc906cf2231b90e8f435f7592bb0301cef72"} Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.308567 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b978c4f56-2cdxx" podStartSLOduration=3.308527505 podStartE2EDuration="3.308527505s" podCreationTimestamp="2025-11-25 10:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:55.304774608 +0000 UTC m=+1152.118391195" watchObservedRunningTime="2025-11-25 10:55:55.308527505 +0000 UTC m=+1152.122144112" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.733352 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.764209 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-swift-storage-0\") pod \"c53b29a3-e64e-459c-a1b8-d700ee54486b\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.766497 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8dcv\" (UniqueName: \"kubernetes.io/projected/c53b29a3-e64e-459c-a1b8-d700ee54486b-kube-api-access-r8dcv\") pod \"c53b29a3-e64e-459c-a1b8-d700ee54486b\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.766687 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-nb\") pod \"c53b29a3-e64e-459c-a1b8-d700ee54486b\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.766725 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-sb\") pod \"c53b29a3-e64e-459c-a1b8-d700ee54486b\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.766769 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-svc\") pod \"c53b29a3-e64e-459c-a1b8-d700ee54486b\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.792252 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c53b29a3-e64e-459c-a1b8-d700ee54486b-kube-api-access-r8dcv" (OuterVolumeSpecName: "kube-api-access-r8dcv") pod "c53b29a3-e64e-459c-a1b8-d700ee54486b" (UID: "c53b29a3-e64e-459c-a1b8-d700ee54486b"). InnerVolumeSpecName "kube-api-access-r8dcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.809120 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c53b29a3-e64e-459c-a1b8-d700ee54486b" (UID: "c53b29a3-e64e-459c-a1b8-d700ee54486b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.829610 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c53b29a3-e64e-459c-a1b8-d700ee54486b" (UID: "c53b29a3-e64e-459c-a1b8-d700ee54486b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.834527 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c599bcb65-hgc2v"] Nov 25 10:55:55 crc kubenswrapper[4696]: E1125 10:55:55.835703 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53b29a3-e64e-459c-a1b8-d700ee54486b" containerName="init" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.835847 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53b29a3-e64e-459c-a1b8-d700ee54486b" containerName="init" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.836149 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c53b29a3-e64e-459c-a1b8-d700ee54486b" containerName="init" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.837308 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.847256 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.847971 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.854236 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c53b29a3-e64e-459c-a1b8-d700ee54486b" (UID: "c53b29a3-e64e-459c-a1b8-d700ee54486b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.858997 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c599bcb65-hgc2v"] Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.881456 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-config\") pod \"c53b29a3-e64e-459c-a1b8-d700ee54486b\" (UID: \"c53b29a3-e64e-459c-a1b8-d700ee54486b\") " Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.882014 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.882034 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.882047 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.882058 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8dcv\" (UniqueName: \"kubernetes.io/projected/c53b29a3-e64e-459c-a1b8-d700ee54486b-kube-api-access-r8dcv\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.937569 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c53b29a3-e64e-459c-a1b8-d700ee54486b" (UID: "c53b29a3-e64e-459c-a1b8-d700ee54486b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.960154 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-config" (OuterVolumeSpecName: "config") pod "c53b29a3-e64e-459c-a1b8-d700ee54486b" (UID: "c53b29a3-e64e-459c-a1b8-d700ee54486b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.983982 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-public-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984256 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-httpd-config\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-ovndb-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984445 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-config\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984549 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-combined-ca-bundle\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984614 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-internal-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984704 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w9hm\" (UniqueName: \"kubernetes.io/projected/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-kube-api-access-6w9hm\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984863 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:55 crc kubenswrapper[4696]: I1125 10:55:55.984931 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53b29a3-e64e-459c-a1b8-d700ee54486b-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.085968 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-public-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.086339 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-httpd-config\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.086899 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-ovndb-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.087016 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-config\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.087130 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-combined-ca-bundle\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.087205 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-internal-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.087316 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w9hm\" (UniqueName: \"kubernetes.io/projected/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-kube-api-access-6w9hm\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.094511 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-httpd-config\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.096637 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-internal-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.096955 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-ovndb-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.098075 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-combined-ca-bundle\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.098238 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-config\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.100464 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-public-tls-certs\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.126322 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w9hm\" (UniqueName: \"kubernetes.io/projected/412cdec0-a9b1-4796-82f6-95c1f6b9b6ae-kube-api-access-6w9hm\") pod \"neutron-6c599bcb65-hgc2v\" (UID: \"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae\") " pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.310682 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.317620 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" event={"ID":"5d51458e-8948-4b70-91ba-02f5a4a323f0","Type":"ContainerStarted","Data":"0aefa896a0e819550b79d8d7ead9800a5b931632d05fd505cbeae33aaeb97a13"} Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.318842 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.326032 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" event={"ID":"c53b29a3-e64e-459c-a1b8-d700ee54486b","Type":"ContainerDied","Data":"7dca901078cb99be9b4f9941f8df086402b7a056b3d868fcaa01fddf05d6c1f5"} Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.326081 4696 scope.go:117] "RemoveContainer" containerID="23501198fceb73df2acc2e09d00dcb299d9596787eb72b296d09c3149386c152" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.326208 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-nkk2q" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.347124 4696 generic.go:334] "Generic (PLEG): container finished" podID="96044f36-992b-44ce-89b4-a059b1efc117" containerID="b9175f8833ce38cbf7619b6d702ee534c161e387b5ab0ec86cc94cf0fc3e71ee" exitCode=0 Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.347195 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerDied","Data":"b9175f8833ce38cbf7619b6d702ee534c161e387b5ab0ec86cc94cf0fc3e71ee"} Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.354031 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-89b7b674-c7w92" event={"ID":"4b6a593d-529f-4d37-a4cb-c86dd239379e","Type":"ContainerStarted","Data":"7876ce0637757831b010ebc24ef4c0498769c92074b4654f7007e235ce7a0c8b"} Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.354069 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.354326 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.358582 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" podStartSLOduration=4.358554112 podStartE2EDuration="4.358554112s" podCreationTimestamp="2025-11-25 10:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:56.34279056 +0000 UTC m=+1153.156407147" watchObservedRunningTime="2025-11-25 10:55:56.358554112 +0000 UTC m=+1153.172170699" Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.395401 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-nkk2q"] Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.401833 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-nkk2q"] Nov 25 10:55:56 crc kubenswrapper[4696]: I1125 10:55:56.448132 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-89b7b674-c7w92" podStartSLOduration=4.448109325 podStartE2EDuration="4.448109325s" podCreationTimestamp="2025-11-25 10:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:56.436740729 +0000 UTC m=+1153.250357316" watchObservedRunningTime="2025-11-25 10:55:56.448109325 +0000 UTC m=+1153.261725912" Nov 25 10:55:57 crc kubenswrapper[4696]: I1125 10:55:57.979459 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.097708 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c53b29a3-e64e-459c-a1b8-d700ee54486b" path="/var/lib/kubelet/pods/c53b29a3-e64e-459c-a1b8-d700ee54486b/volumes" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.140097 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.184178 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.231647 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-config-data\") pod \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.232504 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-db-sync-config-data\") pod \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.233549 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9353d93a-d886-43d0-a4d4-6ad17a28f52a-etc-machine-id\") pod \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.233581 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-scripts\") pod \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.233638 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skp7d\" (UniqueName: \"kubernetes.io/projected/9353d93a-d886-43d0-a4d4-6ad17a28f52a-kube-api-access-skp7d\") pod \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.233680 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-combined-ca-bundle\") pod \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\" (UID: \"9353d93a-d886-43d0-a4d4-6ad17a28f52a\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.242798 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9353d93a-d886-43d0-a4d4-6ad17a28f52a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9353d93a-d886-43d0-a4d4-6ad17a28f52a" (UID: "9353d93a-d886-43d0-a4d4-6ad17a28f52a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.313854 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9353d93a-d886-43d0-a4d4-6ad17a28f52a" (UID: "9353d93a-d886-43d0-a4d4-6ad17a28f52a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.316332 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9353d93a-d886-43d0-a4d4-6ad17a28f52a-kube-api-access-skp7d" (OuterVolumeSpecName: "kube-api-access-skp7d") pod "9353d93a-d886-43d0-a4d4-6ad17a28f52a" (UID: "9353d93a-d886-43d0-a4d4-6ad17a28f52a"). InnerVolumeSpecName "kube-api-access-skp7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.323241 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-scripts" (OuterVolumeSpecName: "scripts") pod "9353d93a-d886-43d0-a4d4-6ad17a28f52a" (UID: "9353d93a-d886-43d0-a4d4-6ad17a28f52a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.324824 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9353d93a-d886-43d0-a4d4-6ad17a28f52a" (UID: "9353d93a-d886-43d0-a4d4-6ad17a28f52a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.336460 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-sg-core-conf-yaml\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.336571 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9bm4\" (UniqueName: \"kubernetes.io/projected/96044f36-992b-44ce-89b4-a059b1efc117-kube-api-access-b9bm4\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.336656 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-scripts\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.341298 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-log-httpd\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.341379 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-config-data\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.341409 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-run-httpd\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.341475 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-combined-ca-bundle\") pod \"96044f36-992b-44ce-89b4-a059b1efc117\" (UID: \"96044f36-992b-44ce-89b4-a059b1efc117\") " Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.342197 4696 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.342223 4696 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9353d93a-d886-43d0-a4d4-6ad17a28f52a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.342235 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.342248 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skp7d\" (UniqueName: \"kubernetes.io/projected/9353d93a-d886-43d0-a4d4-6ad17a28f52a-kube-api-access-skp7d\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.342262 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.349909 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.353865 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.380165 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-scripts" (OuterVolumeSpecName: "scripts") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.394426 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.395215 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96044f36-992b-44ce-89b4-a059b1efc117-kube-api-access-b9bm4" (OuterVolumeSpecName: "kube-api-access-b9bm4") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "kube-api-access-b9bm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.415967 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7qtsq" event={"ID":"9353d93a-d886-43d0-a4d4-6ad17a28f52a","Type":"ContainerDied","Data":"6e1a110e197d7d13f572214aeb6be9ed52da02dec32060e419454586be90639b"} Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.416001 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e1a110e197d7d13f572214aeb6be9ed52da02dec32060e419454586be90639b" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.416047 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7qtsq" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.450531 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.450592 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96044f36-992b-44ce-89b4-a059b1efc117","Type":"ContainerDied","Data":"52d57a14c64d62074b66c6acfebe6d94ab2bc51ac094a524ffdee1324b67f7de"} Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.450658 4696 scope.go:117] "RemoveContainer" containerID="9ed29f07b52672503f9cf758e7c8520c54dd65d393161bd977d1e03aeff4522f" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.453016 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9bm4\" (UniqueName: \"kubernetes.io/projected/96044f36-992b-44ce-89b4-a059b1efc117-kube-api-access-b9bm4\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.453040 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.453049 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.453058 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96044f36-992b-44ce-89b4-a059b1efc117-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.586545 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-config-data" (OuterVolumeSpecName: "config-data") pod "9353d93a-d886-43d0-a4d4-6ad17a28f52a" (UID: "9353d93a-d886-43d0-a4d4-6ad17a28f52a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.634078 4696 scope.go:117] "RemoveContainer" containerID="b2cff43ffe356daf4428055958f7db69c82732c4acd68ddc2c6dafad540cceb5" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.664341 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9353d93a-d886-43d0-a4d4-6ad17a28f52a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.683870 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.712599 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.732883 4696 scope.go:117] "RemoveContainer" containerID="b9175f8833ce38cbf7619b6d702ee534c161e387b5ab0ec86cc94cf0fc3e71ee" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.765536 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.765571 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.769382 4696 scope.go:117] "RemoveContainer" containerID="b129008fc63a532aaeecc84712dd6d46fb99265b4c627542ad08d8514e68301d" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.804430 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c599bcb65-hgc2v"] Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.893362 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-config-data" (OuterVolumeSpecName: "config-data") pod "96044f36-992b-44ce-89b4-a059b1efc117" (UID: "96044f36-992b-44ce-89b4-a059b1efc117"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:58 crc kubenswrapper[4696]: I1125 10:55:58.974843 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96044f36-992b-44ce-89b4-a059b1efc117-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.131383 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.134838 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.151549 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: E1125 10:55:59.152021 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-notification-agent" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152040 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-notification-agent" Nov 25 10:55:59 crc kubenswrapper[4696]: E1125 10:55:59.152053 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="sg-core" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152062 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="sg-core" Nov 25 10:55:59 crc kubenswrapper[4696]: E1125 10:55:59.152076 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="proxy-httpd" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152083 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="proxy-httpd" Nov 25 10:55:59 crc kubenswrapper[4696]: E1125 10:55:59.152128 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" containerName="cinder-db-sync" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152137 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" containerName="cinder-db-sync" Nov 25 10:55:59 crc kubenswrapper[4696]: E1125 10:55:59.152158 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-central-agent" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152166 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-central-agent" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152367 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-central-agent" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152379 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="sg-core" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152395 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="ceilometer-notification-agent" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152410 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" containerName="cinder-db-sync" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.152432 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="96044f36-992b-44ce-89b4-a059b1efc117" containerName="proxy-httpd" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.155421 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.158312 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.158606 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.165796 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.285274 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-config-data\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.285908 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.286619 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-log-httpd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.287062 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-scripts\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.287240 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-run-httpd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.287401 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.287524 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqrjd\" (UniqueName: \"kubernetes.io/projected/b12fabdd-7484-4bcc-b3e5-e583e5742383-kube-api-access-gqrjd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.388996 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-config-data\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.389045 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.389079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-log-httpd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.389127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-scripts\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.389188 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-run-httpd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.389204 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.389249 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqrjd\" (UniqueName: \"kubernetes.io/projected/b12fabdd-7484-4bcc-b3e5-e583e5742383-kube-api-access-gqrjd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.390025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-log-httpd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.392393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-run-httpd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.401395 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.402062 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.407254 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-config-data\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.448787 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-scripts\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.477141 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqrjd\" (UniqueName: \"kubernetes.io/projected/b12fabdd-7484-4bcc-b3e5-e583e5742383-kube-api-access-gqrjd\") pod \"ceilometer-0\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.522744 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.544917 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.547296 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" event={"ID":"9e0f5310-cf60-4125-9b91-2dfda5383c7d","Type":"ContainerStarted","Data":"4922d5aeb351284eef02e74719ac1191be828bb484eb1a80efc94561e88938d1"} Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.547337 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" event={"ID":"9e0f5310-cf60-4125-9b91-2dfda5383c7d","Type":"ContainerStarted","Data":"d1205f7aac3a3d0ae4b80b15c5222cb2ffb2a1122b693eab5b645dcd6167afb0"} Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.547427 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.557378 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.557610 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.557735 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.557872 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jl954" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.586459 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.588924 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64b77754f5-27gpf" event={"ID":"22dbae09-113b-4136-b9d6-6162815b5ef7","Type":"ContainerStarted","Data":"962ff89e17ea2b39aba769665c994062de4ce8256401b9b4f91be66aff83def1"} Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.588962 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64b77754f5-27gpf" event={"ID":"22dbae09-113b-4136-b9d6-6162815b5ef7","Type":"ContainerStarted","Data":"3070620b548117742e9075d8437c40ef66ee7afc39c4cebd451598e3990642f4"} Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.609988 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c599bcb65-hgc2v" event={"ID":"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae","Type":"ContainerStarted","Data":"bc6d8e338ab5ed6490791a9441e170a6d37d096fe7a73880f6042c72380ee193"} Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.610049 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c599bcb65-hgc2v" event={"ID":"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae","Type":"ContainerStarted","Data":"5885244a382451a69e09bca7ff82f9815f4a1c71ebfa098e10b57d8741c1af28"} Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.653288 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-66cc7b58db-xph79" podStartSLOduration=3.54610877 podStartE2EDuration="7.65327187s" podCreationTimestamp="2025-11-25 10:55:52 +0000 UTC" firstStartedPulling="2025-11-25 10:55:53.965620244 +0000 UTC m=+1150.779236831" lastFinishedPulling="2025-11-25 10:55:58.072783344 +0000 UTC m=+1154.886399931" observedRunningTime="2025-11-25 10:55:59.62004262 +0000 UTC m=+1156.433659207" watchObservedRunningTime="2025-11-25 10:55:59.65327187 +0000 UTC m=+1156.466888447" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.727181 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-z7stz"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.727764 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerName="dnsmasq-dns" containerID="cri-o://0aefa896a0e819550b79d8d7ead9800a5b931632d05fd505cbeae33aaeb97a13" gracePeriod=10 Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.732800 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.732846 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.732893 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.733001 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d7d11e20-7b1c-4407-b1d8-391a42e907b2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.733036 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx5dw\" (UniqueName: \"kubernetes.io/projected/d7d11e20-7b1c-4407-b1d8-391a42e907b2-kube-api-access-xx5dw\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.733106 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.775009 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dt2q7"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.776400 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.789042 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-64b77754f5-27gpf" podStartSLOduration=3.967281789 podStartE2EDuration="7.789020338s" podCreationTimestamp="2025-11-25 10:55:52 +0000 UTC" firstStartedPulling="2025-11-25 10:55:54.294773923 +0000 UTC m=+1151.108390510" lastFinishedPulling="2025-11-25 10:55:58.116512482 +0000 UTC m=+1154.930129059" observedRunningTime="2025-11-25 10:55:59.742074968 +0000 UTC m=+1156.555691565" watchObservedRunningTime="2025-11-25 10:55:59.789020338 +0000 UTC m=+1156.602636925" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.832749 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dt2q7"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.834390 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.834431 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.834462 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.834510 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d7d11e20-7b1c-4407-b1d8-391a42e907b2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.834528 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx5dw\" (UniqueName: \"kubernetes.io/projected/d7d11e20-7b1c-4407-b1d8-391a42e907b2-kube-api-access-xx5dw\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.834615 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.841156 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d7d11e20-7b1c-4407-b1d8-391a42e907b2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.866239 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.867320 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.885139 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.887334 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.915293 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx5dw\" (UniqueName: \"kubernetes.io/projected/d7d11e20-7b1c-4407-b1d8-391a42e907b2-kube-api-access-xx5dw\") pod \"cinder-scheduler-0\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " pod="openstack/cinder-scheduler-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.935615 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-config\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.935694 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.935851 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.935999 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htp7x\" (UniqueName: \"kubernetes.io/projected/751aa451-111d-4c91-b4bc-c0d139fa14be-kube-api-access-htp7x\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.936159 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-svc\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.936278 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.943783 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.945306 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.952373 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.965824 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:55:59 crc kubenswrapper[4696]: I1125 10:55:59.982129 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043109 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c1ef208-c90a-458e-a128-e89476c69a33-logs\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043156 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-svc\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043216 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-scripts\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043233 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043268 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-config\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043299 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data-custom\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043469 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043511 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043552 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htp7x\" (UniqueName: \"kubernetes.io/projected/751aa451-111d-4c91-b4bc-c0d139fa14be-kube-api-access-htp7x\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043574 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c1ef208-c90a-458e-a128-e89476c69a33-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043695 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.043736 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r57s4\" (UniqueName: \"kubernetes.io/projected/4c1ef208-c90a-458e-a128-e89476c69a33-kube-api-access-r57s4\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.044680 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-svc\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.045340 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.045891 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-config\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.046396 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.046955 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.139538 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96044f36-992b-44ce-89b4-a059b1efc117" path="/var/lib/kubelet/pods/96044f36-992b-44ce-89b4-a059b1efc117/volumes" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145189 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-scripts\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145252 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145318 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data-custom\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145389 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c1ef208-c90a-458e-a128-e89476c69a33-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145409 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r57s4\" (UniqueName: \"kubernetes.io/projected/4c1ef208-c90a-458e-a128-e89476c69a33-kube-api-access-r57s4\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c1ef208-c90a-458e-a128-e89476c69a33-logs\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.145860 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c1ef208-c90a-458e-a128-e89476c69a33-logs\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.147324 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c1ef208-c90a-458e-a128-e89476c69a33-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.152894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htp7x\" (UniqueName: \"kubernetes.io/projected/751aa451-111d-4c91-b4bc-c0d139fa14be-kube-api-access-htp7x\") pod \"dnsmasq-dns-6578955fd5-dt2q7\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.155169 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-scripts\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.159442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data-custom\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.180750 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.181193 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.217368 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r57s4\" (UniqueName: \"kubernetes.io/projected/4c1ef208-c90a-458e-a128-e89476c69a33-kube-api-access-r57s4\") pod \"cinder-api-0\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.319694 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.405725 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9c4d748fd-g4t2c"] Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.407236 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.410762 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.410952 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.430489 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9c4d748fd-g4t2c"] Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.433129 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.473685 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c86g\" (UniqueName: \"kubernetes.io/projected/c0f652ee-e64b-423b-902a-bab2bd12d46f-kube-api-access-2c86g\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.473917 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-internal-tls-certs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.474015 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-config-data-custom\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.474170 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-public-tls-certs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.474322 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0f652ee-e64b-423b-902a-bab2bd12d46f-logs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.474784 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-config-data\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.474941 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-combined-ca-bundle\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598317 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0f652ee-e64b-423b-902a-bab2bd12d46f-logs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598381 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-config-data\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598523 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-combined-ca-bundle\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598641 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c86g\" (UniqueName: \"kubernetes.io/projected/c0f652ee-e64b-423b-902a-bab2bd12d46f-kube-api-access-2c86g\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598690 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-internal-tls-certs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598733 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-config-data-custom\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.598804 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-public-tls-certs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.602327 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0f652ee-e64b-423b-902a-bab2bd12d46f-logs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.610007 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-combined-ca-bundle\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.634318 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-internal-tls-certs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.634942 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-config-data-custom\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.636790 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-config-data\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.638158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0f652ee-e64b-423b-902a-bab2bd12d46f-public-tls-certs\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.641961 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c86g\" (UniqueName: \"kubernetes.io/projected/c0f652ee-e64b-423b-902a-bab2bd12d46f-kube-api-access-2c86g\") pod \"barbican-api-9c4d748fd-g4t2c\" (UID: \"c0f652ee-e64b-423b-902a-bab2bd12d46f\") " pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.647215 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.718452 4696 generic.go:334] "Generic (PLEG): container finished" podID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerID="0aefa896a0e819550b79d8d7ead9800a5b931632d05fd505cbeae33aaeb97a13" exitCode=0 Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.718529 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" event={"ID":"5d51458e-8948-4b70-91ba-02f5a4a323f0","Type":"ContainerDied","Data":"0aefa896a0e819550b79d8d7ead9800a5b931632d05fd505cbeae33aaeb97a13"} Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.749751 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c599bcb65-hgc2v" event={"ID":"412cdec0-a9b1-4796-82f6-95c1f6b9b6ae","Type":"ContainerStarted","Data":"3b7efdd187ca371d4a0cfeed7ad1e1a09aa3f0e88749aa043721afaed1962e2a"} Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.750751 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.818851 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c599bcb65-hgc2v" podStartSLOduration=5.818824073 podStartE2EDuration="5.818824073s" podCreationTimestamp="2025-11-25 10:55:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:00.814875419 +0000 UTC m=+1157.628492016" watchObservedRunningTime="2025-11-25 10:56:00.818824073 +0000 UTC m=+1157.632440660" Nov 25 10:56:00 crc kubenswrapper[4696]: I1125 10:56:00.902316 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:00 crc kubenswrapper[4696]: W1125 10:56:00.937109 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb12fabdd_7484_4bcc_b3e5_e583e5742383.slice/crio-be0a4297ad834cc8271358c1c78b5465a3aeb15a0f9c6660ea911e10b43f8973 WatchSource:0}: Error finding container be0a4297ad834cc8271358c1c78b5465a3aeb15a0f9c6660ea911e10b43f8973: Status 404 returned error can't find the container with id be0a4297ad834cc8271358c1c78b5465a3aeb15a0f9c6660ea911e10b43f8973 Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.242417 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.478541 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.570569 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.585262 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-sb\") pod \"5d51458e-8948-4b70-91ba-02f5a4a323f0\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.585404 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-config\") pod \"5d51458e-8948-4b70-91ba-02f5a4a323f0\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.585432 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-nb\") pod \"5d51458e-8948-4b70-91ba-02f5a4a323f0\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.585527 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc9xr\" (UniqueName: \"kubernetes.io/projected/5d51458e-8948-4b70-91ba-02f5a4a323f0-kube-api-access-zc9xr\") pod \"5d51458e-8948-4b70-91ba-02f5a4a323f0\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.585546 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-swift-storage-0\") pod \"5d51458e-8948-4b70-91ba-02f5a4a323f0\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.585590 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-svc\") pod \"5d51458e-8948-4b70-91ba-02f5a4a323f0\" (UID: \"5d51458e-8948-4b70-91ba-02f5a4a323f0\") " Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.610497 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d51458e-8948-4b70-91ba-02f5a4a323f0-kube-api-access-zc9xr" (OuterVolumeSpecName: "kube-api-access-zc9xr") pod "5d51458e-8948-4b70-91ba-02f5a4a323f0" (UID: "5d51458e-8948-4b70-91ba-02f5a4a323f0"). InnerVolumeSpecName "kube-api-access-zc9xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.690996 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc9xr\" (UniqueName: \"kubernetes.io/projected/5d51458e-8948-4b70-91ba-02f5a4a323f0-kube-api-access-zc9xr\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.706124 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-config" (OuterVolumeSpecName: "config") pod "5d51458e-8948-4b70-91ba-02f5a4a323f0" (UID: "5d51458e-8948-4b70-91ba-02f5a4a323f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.793343 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.794283 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerStarted","Data":"be0a4297ad834cc8271358c1c78b5465a3aeb15a0f9c6660ea911e10b43f8973"} Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.799240 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" event={"ID":"5d51458e-8948-4b70-91ba-02f5a4a323f0","Type":"ContainerDied","Data":"e8cc8b99d8e21e29dc2081d22cbe189e9a3c2eda2d103713d82a2dbbf5441b54"} Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.799283 4696 scope.go:117] "RemoveContainer" containerID="0aefa896a0e819550b79d8d7ead9800a5b931632d05fd505cbeae33aaeb97a13" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.799403 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-z7stz" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.812469 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4c1ef208-c90a-458e-a128-e89476c69a33","Type":"ContainerStarted","Data":"13624b60027b04bfb610e4780064580abbfb8cf8091eb4156c281daf441964b1"} Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.830621 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d7d11e20-7b1c-4407-b1d8-391a42e907b2","Type":"ContainerStarted","Data":"53c1fc835749643c1c72b3f60b0da8ef9f4219f588cf82a9df53ad735be43408"} Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.857673 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9c4d748fd-g4t2c"] Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.902351 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5d51458e-8948-4b70-91ba-02f5a4a323f0" (UID: "5d51458e-8948-4b70-91ba-02f5a4a323f0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.909211 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d51458e-8948-4b70-91ba-02f5a4a323f0" (UID: "5d51458e-8948-4b70-91ba-02f5a4a323f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.918323 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5d51458e-8948-4b70-91ba-02f5a4a323f0" (UID: "5d51458e-8948-4b70-91ba-02f5a4a323f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.974301 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5d51458e-8948-4b70-91ba-02f5a4a323f0" (UID: "5d51458e-8948-4b70-91ba-02f5a4a323f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.996127 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.996158 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.996170 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:01 crc kubenswrapper[4696]: I1125 10:56:01.996179 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d51458e-8948-4b70-91ba-02f5a4a323f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.011755 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dt2q7"] Nov 25 10:56:02 crc kubenswrapper[4696]: W1125 10:56:02.066019 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod751aa451_111d_4c91_b4bc_c0d139fa14be.slice/crio-b2fb44b083cd716fb074819c2eee1ae4136b225cf13bf9770753e0c9c8d23673 WatchSource:0}: Error finding container b2fb44b083cd716fb074819c2eee1ae4136b225cf13bf9770753e0c9c8d23673: Status 404 returned error can't find the container with id b2fb44b083cd716fb074819c2eee1ae4136b225cf13bf9770753e0c9c8d23673 Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.077215 4696 scope.go:117] "RemoveContainer" containerID="40c05b71aa3b47f585a3d1d2dbe93e36b4f218b2c4967de6786b21b251f9941e" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.328383 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-z7stz"] Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.351510 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-z7stz"] Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.778926 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.779243 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.781097 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"b743c34bdb2c734c6424f22dad595eb14070e34899fbea0e6b459832b98e3148"} pod="openstack/horizon-7c8fd89c8d-5vb5f" containerMessage="Container horizon failed startup probe, will be restarted" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.781141 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" containerID="cri-o://b743c34bdb2c734c6424f22dad595eb14070e34899fbea0e6b459832b98e3148" gracePeriod=30 Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.861517 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9c4d748fd-g4t2c" event={"ID":"c0f652ee-e64b-423b-902a-bab2bd12d46f","Type":"ContainerStarted","Data":"0ddeb819049883a3e7e9b1e0857671a9188ba3f90b00a5d04d3cd652aa5f5abd"} Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.861561 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9c4d748fd-g4t2c" event={"ID":"c0f652ee-e64b-423b-902a-bab2bd12d46f","Type":"ContainerStarted","Data":"13a45d899d801c4c75ad26ac94325f16980a40a81452f44129897fa43cfe11f9"} Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.872628 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerStarted","Data":"6886bfd17a01518b9a498155f6ec3195289c3636b9a72d054ddc5f181733e425"} Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.879574 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6558fc797b-2qp4r" podUID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.879641 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.880417 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"8ac4f70641365c40d7d066cd108525fd104f0eaf62202401f21b85c5a5995efc"} pod="openstack/horizon-6558fc797b-2qp4r" containerMessage="Container horizon failed startup probe, will be restarted" Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.880459 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6558fc797b-2qp4r" podUID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerName="horizon" containerID="cri-o://8ac4f70641365c40d7d066cd108525fd104f0eaf62202401f21b85c5a5995efc" gracePeriod=30 Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.882564 4696 generic.go:334] "Generic (PLEG): container finished" podID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerID="7ff9ae8b9c72802068a1f15a4ae0531f3629c49034c83f8f70e0ddfc34f3a7f7" exitCode=0 Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.884119 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" event={"ID":"751aa451-111d-4c91-b4bc-c0d139fa14be","Type":"ContainerDied","Data":"7ff9ae8b9c72802068a1f15a4ae0531f3629c49034c83f8f70e0ddfc34f3a7f7"} Nov 25 10:56:02 crc kubenswrapper[4696]: I1125 10:56:02.884167 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" event={"ID":"751aa451-111d-4c91-b4bc-c0d139fa14be","Type":"ContainerStarted","Data":"b2fb44b083cd716fb074819c2eee1ae4136b225cf13bf9770753e0c9c8d23673"} Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.906650 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" event={"ID":"751aa451-111d-4c91-b4bc-c0d139fa14be","Type":"ContainerStarted","Data":"5657b9c21cf0b28b8f20a3b0bcac6c9ab2367ea6e314788a44a66455cdccb4ae"} Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.908031 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.938820 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9c4d748fd-g4t2c" event={"ID":"c0f652ee-e64b-423b-902a-bab2bd12d46f","Type":"ContainerStarted","Data":"a5f066686ea51190866b0c3fb2ad9734405bf3b0c69d1476bd77812dd2fa0f9d"} Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.939563 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.939592 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.989177 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerStarted","Data":"ad986f8b09ab9b48845583ca4d819f7c809fdc72a205248b7a96988b2ad37292"} Nov 25 10:56:03 crc kubenswrapper[4696]: I1125 10:56:03.997094 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4c1ef208-c90a-458e-a128-e89476c69a33","Type":"ContainerStarted","Data":"1f854fc5e2148b76839c7980b26081f0bdf9f9141e2bec1987935c777037ba23"} Nov 25 10:56:04 crc kubenswrapper[4696]: I1125 10:56:04.002825 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d7d11e20-7b1c-4407-b1d8-391a42e907b2","Type":"ContainerStarted","Data":"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4"} Nov 25 10:56:04 crc kubenswrapper[4696]: I1125 10:56:04.006820 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" podStartSLOduration=5.00680215 podStartE2EDuration="5.00680215s" podCreationTimestamp="2025-11-25 10:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:03.989700285 +0000 UTC m=+1160.803316872" watchObservedRunningTime="2025-11-25 10:56:04.00680215 +0000 UTC m=+1160.820418737" Nov 25 10:56:04 crc kubenswrapper[4696]: I1125 10:56:04.076338 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" path="/var/lib/kubelet/pods/5d51458e-8948-4b70-91ba-02f5a4a323f0/volumes" Nov 25 10:56:04 crc kubenswrapper[4696]: I1125 10:56:04.085554 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9c4d748fd-g4t2c" podStartSLOduration=4.085533094 podStartE2EDuration="4.085533094s" podCreationTimestamp="2025-11-25 10:56:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:04.054031508 +0000 UTC m=+1160.867648095" watchObservedRunningTime="2025-11-25 10:56:04.085533094 +0000 UTC m=+1160.899149681" Nov 25 10:56:04 crc kubenswrapper[4696]: I1125 10:56:04.405527 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:05 crc kubenswrapper[4696]: I1125 10:56:05.028734 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d7d11e20-7b1c-4407-b1d8-391a42e907b2","Type":"ContainerStarted","Data":"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582"} Nov 25 10:56:05 crc kubenswrapper[4696]: I1125 10:56:05.057354 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.970879626 podStartE2EDuration="6.057334073s" podCreationTimestamp="2025-11-25 10:55:59 +0000 UTC" firstStartedPulling="2025-11-25 10:56:01.251531942 +0000 UTC m=+1158.065148529" lastFinishedPulling="2025-11-25 10:56:02.337986389 +0000 UTC m=+1159.151602976" observedRunningTime="2025-11-25 10:56:05.048750975 +0000 UTC m=+1161.862367572" watchObservedRunningTime="2025-11-25 10:56:05.057334073 +0000 UTC m=+1161.870950660" Nov 25 10:56:06 crc kubenswrapper[4696]: I1125 10:56:06.039178 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerStarted","Data":"a67648e6342409a1ea3448b385e39d5c3f42598c4383f1fddbf65670b580c176"} Nov 25 10:56:06 crc kubenswrapper[4696]: I1125 10:56:06.041478 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api-log" containerID="cri-o://1f854fc5e2148b76839c7980b26081f0bdf9f9141e2bec1987935c777037ba23" gracePeriod=30 Nov 25 10:56:06 crc kubenswrapper[4696]: I1125 10:56:06.041585 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api" containerID="cri-o://f701dbc71200da9e920f1193f1ded1c9dc3ffd72ba0a90f6105be22e474e00aa" gracePeriod=30 Nov 25 10:56:06 crc kubenswrapper[4696]: I1125 10:56:06.041894 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4c1ef208-c90a-458e-a128-e89476c69a33","Type":"ContainerStarted","Data":"f701dbc71200da9e920f1193f1ded1c9dc3ffd72ba0a90f6105be22e474e00aa"} Nov 25 10:56:06 crc kubenswrapper[4696]: I1125 10:56:06.042516 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:56:06 crc kubenswrapper[4696]: I1125 10:56:06.066766 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.066752639 podStartE2EDuration="7.066752639s" podCreationTimestamp="2025-11-25 10:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:06.064398686 +0000 UTC m=+1162.878015273" watchObservedRunningTime="2025-11-25 10:56:06.066752639 +0000 UTC m=+1162.880369226" Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.051796 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerStarted","Data":"90dd0e2ba15b7ad04da0f23f9e8df06a2f527c6cabed6095b200234562d07b7f"} Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.053209 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.053907 4696 generic.go:334] "Generic (PLEG): container finished" podID="4c1ef208-c90a-458e-a128-e89476c69a33" containerID="1f854fc5e2148b76839c7980b26081f0bdf9f9141e2bec1987935c777037ba23" exitCode=143 Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.053942 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4c1ef208-c90a-458e-a128-e89476c69a33","Type":"ContainerDied","Data":"1f854fc5e2148b76839c7980b26081f0bdf9f9141e2bec1987935c777037ba23"} Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.078784 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.436154199 podStartE2EDuration="8.078747367s" podCreationTimestamp="2025-11-25 10:55:59 +0000 UTC" firstStartedPulling="2025-11-25 10:56:00.957055387 +0000 UTC m=+1157.770671974" lastFinishedPulling="2025-11-25 10:56:06.599648555 +0000 UTC m=+1163.413265142" observedRunningTime="2025-11-25 10:56:07.077936221 +0000 UTC m=+1163.891552808" watchObservedRunningTime="2025-11-25 10:56:07.078747367 +0000 UTC m=+1163.892363954" Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.786983 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:07 crc kubenswrapper[4696]: I1125 10:56:07.787022 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:08 crc kubenswrapper[4696]: I1125 10:56:08.790785 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:08 crc kubenswrapper[4696]: I1125 10:56:08.790986 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:09 crc kubenswrapper[4696]: I1125 10:56:09.983030 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 10:56:09 crc kubenswrapper[4696]: I1125 10:56:09.985812 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.162:8080/\": dial tcp 10.217.0.162:8080: connect: connection refused" Nov 25 10:56:10 crc kubenswrapper[4696]: I1125 10:56:10.434920 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:56:10 crc kubenswrapper[4696]: I1125 10:56:10.550351 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-f8bqt"] Nov 25 10:56:10 crc kubenswrapper[4696]: I1125 10:56:10.550559 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerName="dnsmasq-dns" containerID="cri-o://82ab43d8f8da99faf8b849727008c587b3003c2fd55a87177fed8af175189d52" gracePeriod=10 Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.102979 4696 generic.go:334] "Generic (PLEG): container finished" podID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerID="82ab43d8f8da99faf8b849727008c587b3003c2fd55a87177fed8af175189d52" exitCode=0 Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.103075 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" event={"ID":"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a","Type":"ContainerDied","Data":"82ab43d8f8da99faf8b849727008c587b3003c2fd55a87177fed8af175189d52"} Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.339015 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.373801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-sb\") pod \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.373867 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-swift-storage-0\") pod \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.373963 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-nb\") pod \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.374003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-config\") pod \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.374049 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-svc\") pod \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.374091 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65pnx\" (UniqueName: \"kubernetes.io/projected/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-kube-api-access-65pnx\") pod \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\" (UID: \"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a\") " Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.386129 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-kube-api-access-65pnx" (OuterVolumeSpecName: "kube-api-access-65pnx") pod "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" (UID: "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a"). InnerVolumeSpecName "kube-api-access-65pnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.484022 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65pnx\" (UniqueName: \"kubernetes.io/projected/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-kube-api-access-65pnx\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.560467 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" (UID: "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.579288 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" (UID: "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.583022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-config" (OuterVolumeSpecName: "config") pod "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" (UID: "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.587751 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.587782 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.587791 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.590436 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" (UID: "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.602554 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" (UID: "7c3a8e43-bb23-4fb2-ab40-e2ff6676212a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.689278 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:11 crc kubenswrapper[4696]: I1125 10:56:11.689318 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.117386 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" event={"ID":"7c3a8e43-bb23-4fb2-ab40-e2ff6676212a","Type":"ContainerDied","Data":"d39ddd8d56175f02d7ef3b7fae3609b2f725339d44ab57f250f0577009470f4a"} Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.117437 4696 scope.go:117] "RemoveContainer" containerID="82ab43d8f8da99faf8b849727008c587b3003c2fd55a87177fed8af175189d52" Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.117560 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-f8bqt" Nov 25 10:56:12 crc kubenswrapper[4696]: E1125 10:56:12.147582 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c3a8e43_bb23_4fb2_ab40_e2ff6676212a.slice/crio-d39ddd8d56175f02d7ef3b7fae3609b2f725339d44ab57f250f0577009470f4a\": RecentStats: unable to find data in memory cache]" Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.159859 4696 scope.go:117] "RemoveContainer" containerID="27c33a1491b2001d59ab6a25a63b79c0bc5160c423bf405f0d1b1bd59a049496" Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.198135 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-f8bqt"] Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.229281 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-f8bqt"] Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.461697 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.870846 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:12 crc kubenswrapper[4696]: I1125 10:56:12.870858 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:13 crc kubenswrapper[4696]: I1125 10:56:13.010687 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:56:14 crc kubenswrapper[4696]: I1125 10:56:14.051435 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" path="/var/lib/kubelet/pods/7c3a8e43-bb23-4fb2-ab40-e2ff6676212a/volumes" Nov 25 10:56:14 crc kubenswrapper[4696]: I1125 10:56:14.655951 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-9c4d748fd-g4t2c" podUID="c0f652ee-e64b-423b-902a-bab2bd12d46f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:14 crc kubenswrapper[4696]: I1125 10:56:14.804108 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:56:14 crc kubenswrapper[4696]: I1125 10:56:14.870293 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-646fccfb54-pbhr5" Nov 25 10:56:15 crc kubenswrapper[4696]: I1125 10:56:15.361876 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.164:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:15 crc kubenswrapper[4696]: I1125 10:56:15.552107 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 10:56:15 crc kubenswrapper[4696]: I1125 10:56:15.620119 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:15 crc kubenswrapper[4696]: I1125 10:56:15.633693 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:15 crc kubenswrapper[4696]: I1125 10:56:15.651897 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9c4d748fd-g4t2c" podUID="c0f652ee-e64b-423b-902a-bab2bd12d46f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:16 crc kubenswrapper[4696]: I1125 10:56:16.107881 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5547bcb8fc-b65gm" Nov 25 10:56:16 crc kubenswrapper[4696]: I1125 10:56:16.150777 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="cinder-scheduler" containerID="cri-o://c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4" gracePeriod=30 Nov 25 10:56:16 crc kubenswrapper[4696]: I1125 10:56:16.151306 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="probe" containerID="cri-o://703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582" gracePeriod=30 Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.823734 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.905649 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data-custom\") pod \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.906935 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-combined-ca-bundle\") pod \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.907076 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data\") pod \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.907164 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-scripts\") pod \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.907294 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx5dw\" (UniqueName: \"kubernetes.io/projected/d7d11e20-7b1c-4407-b1d8-391a42e907b2-kube-api-access-xx5dw\") pod \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.907383 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d7d11e20-7b1c-4407-b1d8-391a42e907b2-etc-machine-id\") pod \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\" (UID: \"d7d11e20-7b1c-4407-b1d8-391a42e907b2\") " Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.907900 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7d11e20-7b1c-4407-b1d8-391a42e907b2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d7d11e20-7b1c-4407-b1d8-391a42e907b2" (UID: "d7d11e20-7b1c-4407-b1d8-391a42e907b2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.916548 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d7d11e20-7b1c-4407-b1d8-391a42e907b2" (UID: "d7d11e20-7b1c-4407-b1d8-391a42e907b2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.928839 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d11e20-7b1c-4407-b1d8-391a42e907b2-kube-api-access-xx5dw" (OuterVolumeSpecName: "kube-api-access-xx5dw") pod "d7d11e20-7b1c-4407-b1d8-391a42e907b2" (UID: "d7d11e20-7b1c-4407-b1d8-391a42e907b2"). InnerVolumeSpecName "kube-api-access-xx5dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:17 crc kubenswrapper[4696]: I1125 10:56:17.928927 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-scripts" (OuterVolumeSpecName: "scripts") pod "d7d11e20-7b1c-4407-b1d8-391a42e907b2" (UID: "d7d11e20-7b1c-4407-b1d8-391a42e907b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.008937 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx5dw\" (UniqueName: \"kubernetes.io/projected/d7d11e20-7b1c-4407-b1d8-391a42e907b2-kube-api-access-xx5dw\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.008964 4696 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d7d11e20-7b1c-4407-b1d8-391a42e907b2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.008973 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.008981 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.016382 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7d11e20-7b1c-4407-b1d8-391a42e907b2" (UID: "d7d11e20-7b1c-4407-b1d8-391a42e907b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.093985 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data" (OuterVolumeSpecName: "config-data") pod "d7d11e20-7b1c-4407-b1d8-391a42e907b2" (UID: "d7d11e20-7b1c-4407-b1d8-391a42e907b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.111433 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.111471 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d11e20-7b1c-4407-b1d8-391a42e907b2-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167284 4696 generic.go:334] "Generic (PLEG): container finished" podID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerID="703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582" exitCode=0 Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167312 4696 generic.go:334] "Generic (PLEG): container finished" podID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerID="c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4" exitCode=0 Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167334 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d7d11e20-7b1c-4407-b1d8-391a42e907b2","Type":"ContainerDied","Data":"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582"} Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167373 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d7d11e20-7b1c-4407-b1d8-391a42e907b2","Type":"ContainerDied","Data":"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4"} Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167384 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d7d11e20-7b1c-4407-b1d8-391a42e907b2","Type":"ContainerDied","Data":"53c1fc835749643c1c72b3f60b0da8ef9f4219f588cf82a9df53ad735be43408"} Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167399 4696 scope.go:117] "RemoveContainer" containerID="703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.167790 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.242092 4696 scope.go:117] "RemoveContainer" containerID="c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.250881 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.265015 4696 scope.go:117] "RemoveContainer" containerID="703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.265562 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582\": container with ID starting with 703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582 not found: ID does not exist" containerID="703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.265682 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582"} err="failed to get container status \"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582\": rpc error: code = NotFound desc = could not find container \"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582\": container with ID starting with 703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582 not found: ID does not exist" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.265761 4696 scope.go:117] "RemoveContainer" containerID="c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.266106 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4\": container with ID starting with c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4 not found: ID does not exist" containerID="c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.266190 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4"} err="failed to get container status \"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4\": rpc error: code = NotFound desc = could not find container \"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4\": container with ID starting with c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4 not found: ID does not exist" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.266268 4696 scope.go:117] "RemoveContainer" containerID="703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.266438 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.266686 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582"} err="failed to get container status \"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582\": rpc error: code = NotFound desc = could not find container \"703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582\": container with ID starting with 703b693dd597d573f5e0e1a713dbf7f59d79dac8c3ab36fb216f71c469eec582 not found: ID does not exist" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.266769 4696 scope.go:117] "RemoveContainer" containerID="c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.267183 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4"} err="failed to get container status \"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4\": rpc error: code = NotFound desc = could not find container \"c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4\": container with ID starting with c8ab1be17ae6cad364eac385521131462d168b5bc079a489637c5611015a8cf4 not found: ID does not exist" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.277957 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.278337 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerName="init" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278361 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerName="init" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.278379 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerName="dnsmasq-dns" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278387 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerName="dnsmasq-dns" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.278398 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerName="init" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278405 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerName="init" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.278417 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerName="dnsmasq-dns" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278424 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerName="dnsmasq-dns" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.278438 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="cinder-scheduler" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278447 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="cinder-scheduler" Nov 25 10:56:18 crc kubenswrapper[4696]: E1125 10:56:18.278464 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="probe" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278470 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="probe" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278654 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3a8e43-bb23-4fb2-ab40-e2ff6676212a" containerName="dnsmasq-dns" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278696 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="cinder-scheduler" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278710 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" containerName="probe" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.278723 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d51458e-8948-4b70-91ba-02f5a4a323f0" containerName="dnsmasq-dns" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.286910 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.291060 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.298465 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.423952 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjgs\" (UniqueName: \"kubernetes.io/projected/dccb9177-738d-433c-a2ec-99131e690f89-kube-api-access-4jjgs\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.424030 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dccb9177-738d-433c-a2ec-99131e690f89-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.424073 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-scripts\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.424101 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.424117 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.424134 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-config-data\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.525843 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjgs\" (UniqueName: \"kubernetes.io/projected/dccb9177-738d-433c-a2ec-99131e690f89-kube-api-access-4jjgs\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.525935 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dccb9177-738d-433c-a2ec-99131e690f89-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.525979 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-scripts\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.526007 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.526025 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.526044 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-config-data\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.526989 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dccb9177-738d-433c-a2ec-99131e690f89-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.534456 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-scripts\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.534748 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-config-data\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.536283 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.542861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccb9177-738d-433c-a2ec-99131e690f89-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.556926 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjgs\" (UniqueName: \"kubernetes.io/projected/dccb9177-738d-433c-a2ec-99131e690f89-kube-api-access-4jjgs\") pod \"cinder-scheduler-0\" (UID: \"dccb9177-738d-433c-a2ec-99131e690f89\") " pod="openstack/cinder-scheduler-0" Nov 25 10:56:18 crc kubenswrapper[4696]: I1125 10:56:18.609263 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.331444 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.638232 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.809563 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.810655 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.822477 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6drbh" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.824238 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.824367 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.844440 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.881156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kr7q\" (UniqueName: \"kubernetes.io/projected/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-kube-api-access-2kr7q\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.881278 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.881359 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.881377 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-openstack-config\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.984387 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.984431 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-openstack-config\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.984496 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kr7q\" (UniqueName: \"kubernetes.io/projected/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-kube-api-access-2kr7q\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.984563 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.987214 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-openstack-config\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.993097 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:19 crc kubenswrapper[4696]: I1125 10:56:19.993546 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.015228 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kr7q\" (UniqueName: \"kubernetes.io/projected/14829ad2-5154-4bc3-ac0d-8343c93c0f6d-kube-api-access-2kr7q\") pod \"openstackclient\" (UID: \"14829ad2-5154-4bc3-ac0d-8343c93c0f6d\") " pod="openstack/openstackclient" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.068423 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d11e20-7b1c-4407-b1d8-391a42e907b2" path="/var/lib/kubelet/pods/d7d11e20-7b1c-4407-b1d8-391a42e907b2/volumes" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.179294 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.205080 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dccb9177-738d-433c-a2ec-99131e690f89","Type":"ContainerStarted","Data":"95eba82a58411af3c3fa7b237b8ef0bb18c6d02d238703cb653678983ccf47a0"} Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.664919 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9c4d748fd-g4t2c" podUID="c0f652ee-e64b-423b-902a-bab2bd12d46f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.665858 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9c4d748fd-g4t2c" podUID="c0f652ee-e64b-423b-902a-bab2bd12d46f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.688020 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9c4d748fd-g4t2c" Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.751722 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-89b7b674-c7w92"] Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.751952 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" containerID="cri-o://d1046fa1331aef9c0bf81217545c4717672a0b12a31fa57d7362160831fb9366" gracePeriod=30 Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.752189 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" containerID="cri-o://7876ce0637757831b010ebc24ef4c0498769c92074b4654f7007e235ce7a0c8b" gracePeriod=30 Nov 25 10:56:20 crc kubenswrapper[4696]: I1125 10:56:20.878876 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:56:20 crc kubenswrapper[4696]: W1125 10:56:20.917935 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14829ad2_5154_4bc3_ac0d_8343c93c0f6d.slice/crio-c9c121a517227c42d868ae4aa5b4a15adac38612e334dc5ac100c5a2d07dc3a0 WatchSource:0}: Error finding container c9c121a517227c42d868ae4aa5b4a15adac38612e334dc5ac100c5a2d07dc3a0: Status 404 returned error can't find the container with id c9c121a517227c42d868ae4aa5b4a15adac38612e334dc5ac100c5a2d07dc3a0 Nov 25 10:56:21 crc kubenswrapper[4696]: I1125 10:56:21.218208 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"14829ad2-5154-4bc3-ac0d-8343c93c0f6d","Type":"ContainerStarted","Data":"c9c121a517227c42d868ae4aa5b4a15adac38612e334dc5ac100c5a2d07dc3a0"} Nov 25 10:56:21 crc kubenswrapper[4696]: I1125 10:56:21.219684 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dccb9177-738d-433c-a2ec-99131e690f89","Type":"ContainerStarted","Data":"d9c20f68d45cbfc07188dbb535fe72d959b0306d71fa27164c6b332f78020392"} Nov 25 10:56:21 crc kubenswrapper[4696]: I1125 10:56:21.222347 4696 generic.go:334] "Generic (PLEG): container finished" podID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerID="d1046fa1331aef9c0bf81217545c4717672a0b12a31fa57d7362160831fb9366" exitCode=143 Nov 25 10:56:21 crc kubenswrapper[4696]: I1125 10:56:21.222390 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-89b7b674-c7w92" event={"ID":"4b6a593d-529f-4d37-a4cb-c86dd239379e","Type":"ContainerDied","Data":"d1046fa1331aef9c0bf81217545c4717672a0b12a31fa57d7362160831fb9366"} Nov 25 10:56:22 crc kubenswrapper[4696]: I1125 10:56:22.257813 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dccb9177-738d-433c-a2ec-99131e690f89","Type":"ContainerStarted","Data":"6dd57017ff1c1f4073eb9eee28dae6c98867d32a1953cc542034e037f98b06a5"} Nov 25 10:56:22 crc kubenswrapper[4696]: I1125 10:56:22.282191 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.282172571 podStartE2EDuration="4.282172571s" podCreationTimestamp="2025-11-25 10:56:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:22.278927507 +0000 UTC m=+1179.092544114" watchObservedRunningTime="2025-11-25 10:56:22.282172571 +0000 UTC m=+1179.095789158" Nov 25 10:56:22 crc kubenswrapper[4696]: I1125 10:56:22.701064 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:56:23 crc kubenswrapper[4696]: I1125 10:56:23.609468 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 10:56:23 crc kubenswrapper[4696]: I1125 10:56:23.958216 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:48352->10.217.0.158:9311: read: connection reset by peer" Nov 25 10:56:23 crc kubenswrapper[4696]: I1125 10:56:23.958593 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-89b7b674-c7w92" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:48362->10.217.0.158:9311: read: connection reset by peer" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.281717 4696 generic.go:334] "Generic (PLEG): container finished" podID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerID="7876ce0637757831b010ebc24ef4c0498769c92074b4654f7007e235ce7a0c8b" exitCode=0 Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.283405 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-89b7b674-c7w92" event={"ID":"4b6a593d-529f-4d37-a4cb-c86dd239379e","Type":"ContainerDied","Data":"7876ce0637757831b010ebc24ef4c0498769c92074b4654f7007e235ce7a0c8b"} Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.415825 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.583660 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data\") pod \"4b6a593d-529f-4d37-a4cb-c86dd239379e\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.583794 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data-custom\") pod \"4b6a593d-529f-4d37-a4cb-c86dd239379e\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.583907 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-combined-ca-bundle\") pod \"4b6a593d-529f-4d37-a4cb-c86dd239379e\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.583937 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmbr8\" (UniqueName: \"kubernetes.io/projected/4b6a593d-529f-4d37-a4cb-c86dd239379e-kube-api-access-lmbr8\") pod \"4b6a593d-529f-4d37-a4cb-c86dd239379e\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.584069 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b6a593d-529f-4d37-a4cb-c86dd239379e-logs\") pod \"4b6a593d-529f-4d37-a4cb-c86dd239379e\" (UID: \"4b6a593d-529f-4d37-a4cb-c86dd239379e\") " Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.584902 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b6a593d-529f-4d37-a4cb-c86dd239379e-logs" (OuterVolumeSpecName: "logs") pod "4b6a593d-529f-4d37-a4cb-c86dd239379e" (UID: "4b6a593d-529f-4d37-a4cb-c86dd239379e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.591739 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4b6a593d-529f-4d37-a4cb-c86dd239379e" (UID: "4b6a593d-529f-4d37-a4cb-c86dd239379e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.607433 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b6a593d-529f-4d37-a4cb-c86dd239379e-kube-api-access-lmbr8" (OuterVolumeSpecName: "kube-api-access-lmbr8") pod "4b6a593d-529f-4d37-a4cb-c86dd239379e" (UID: "4b6a593d-529f-4d37-a4cb-c86dd239379e"). InnerVolumeSpecName "kube-api-access-lmbr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.623083 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b6a593d-529f-4d37-a4cb-c86dd239379e" (UID: "4b6a593d-529f-4d37-a4cb-c86dd239379e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.644620 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data" (OuterVolumeSpecName: "config-data") pod "4b6a593d-529f-4d37-a4cb-c86dd239379e" (UID: "4b6a593d-529f-4d37-a4cb-c86dd239379e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.685849 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.685889 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.685899 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b6a593d-529f-4d37-a4cb-c86dd239379e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.685908 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmbr8\" (UniqueName: \"kubernetes.io/projected/4b6a593d-529f-4d37-a4cb-c86dd239379e-kube-api-access-lmbr8\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:24 crc kubenswrapper[4696]: I1125 10:56:24.685917 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b6a593d-529f-4d37-a4cb-c86dd239379e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.333166 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-78446968f5-84c89"] Nov 25 10:56:25 crc kubenswrapper[4696]: E1125 10:56:25.334172 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.334257 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" Nov 25 10:56:25 crc kubenswrapper[4696]: E1125 10:56:25.334365 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.334431 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.334846 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.334948 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" containerName="barbican-api-log" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.336569 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.340205 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.340398 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.340491 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-89b7b674-c7w92" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.340416 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-89b7b674-c7w92" event={"ID":"4b6a593d-529f-4d37-a4cb-c86dd239379e","Type":"ContainerDied","Data":"f17836c73f8efa8bf62315e96d8965ffd5e297ca65fb7ba4cd3b05afe9272d6e"} Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.340763 4696 scope.go:117] "RemoveContainer" containerID="7876ce0637757831b010ebc24ef4c0498769c92074b4654f7007e235ce7a0c8b" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.351480 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.402288 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78446968f5-84c89"] Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.443036 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-89b7b674-c7w92"] Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.443900 4696 scope.go:117] "RemoveContainer" containerID="d1046fa1331aef9c0bf81217545c4717672a0b12a31fa57d7362160831fb9366" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.450438 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-89b7b674-c7w92"] Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.515809 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142ffa20-cd2a-4b6b-90ff-2770e6eed020-log-httpd\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.515865 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-config-data\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.515942 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-public-tls-certs\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.516011 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-combined-ca-bundle\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.516047 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142ffa20-cd2a-4b6b-90ff-2770e6eed020-run-httpd\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.516068 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/142ffa20-cd2a-4b6b-90ff-2770e6eed020-etc-swift\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.516083 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-internal-tls-certs\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.516156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnfss\" (UniqueName: \"kubernetes.io/projected/142ffa20-cd2a-4b6b-90ff-2770e6eed020-kube-api-access-qnfss\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617406 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-combined-ca-bundle\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617464 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142ffa20-cd2a-4b6b-90ff-2770e6eed020-run-httpd\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617488 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/142ffa20-cd2a-4b6b-90ff-2770e6eed020-etc-swift\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617503 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-internal-tls-certs\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617548 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnfss\" (UniqueName: \"kubernetes.io/projected/142ffa20-cd2a-4b6b-90ff-2770e6eed020-kube-api-access-qnfss\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617611 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142ffa20-cd2a-4b6b-90ff-2770e6eed020-log-httpd\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617631 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-config-data\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-public-tls-certs\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.617974 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142ffa20-cd2a-4b6b-90ff-2770e6eed020-run-httpd\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.618463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142ffa20-cd2a-4b6b-90ff-2770e6eed020-log-httpd\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.622647 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-public-tls-certs\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.630601 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/142ffa20-cd2a-4b6b-90ff-2770e6eed020-etc-swift\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.632854 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-internal-tls-certs\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.636197 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnfss\" (UniqueName: \"kubernetes.io/projected/142ffa20-cd2a-4b6b-90ff-2770e6eed020-kube-api-access-qnfss\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.636597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-config-data\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.637151 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ffa20-cd2a-4b6b-90ff-2770e6eed020-combined-ca-bundle\") pod \"swift-proxy-78446968f5-84c89\" (UID: \"142ffa20-cd2a-4b6b-90ff-2770e6eed020\") " pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:25 crc kubenswrapper[4696]: I1125 10:56:25.671713 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.058789 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b6a593d-529f-4d37-a4cb-c86dd239379e" path="/var/lib/kubelet/pods/4b6a593d-529f-4d37-a4cb-c86dd239379e/volumes" Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.327739 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78446968f5-84c89"] Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.340800 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c599bcb65-hgc2v" Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.356083 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78446968f5-84c89" event={"ID":"142ffa20-cd2a-4b6b-90ff-2770e6eed020","Type":"ContainerStarted","Data":"e2dd26f51a5caf57ada5f78292c0313522a3aa4185693871ae66df3b06250484"} Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.479864 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b978c4f56-2cdxx"] Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.480344 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b978c4f56-2cdxx" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-api" containerID="cri-o://88a9999603fdc26e8442e67101cb67c0c71a56b11ac4283bd258f1aaf40ea1d6" gracePeriod=30 Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.480487 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b978c4f56-2cdxx" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-httpd" containerID="cri-o://87d42e215c695a26e3b7b4ca984c8c2baee37d35456521565a5e09c21e03f574" gracePeriod=30 Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.814737 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.815303 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-central-agent" containerID="cri-o://6886bfd17a01518b9a498155f6ec3195289c3636b9a72d054ddc5f181733e425" gracePeriod=30 Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.815811 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="proxy-httpd" containerID="cri-o://90dd0e2ba15b7ad04da0f23f9e8df06a2f527c6cabed6095b200234562d07b7f" gracePeriod=30 Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.815896 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-notification-agent" containerID="cri-o://ad986f8b09ab9b48845583ca4d819f7c809fdc72a205248b7a96988b2ad37292" gracePeriod=30 Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.815950 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="sg-core" containerID="cri-o://a67648e6342409a1ea3448b385e39d5c3f42598c4383f1fddbf65670b580c176" gracePeriod=30 Nov 25 10:56:26 crc kubenswrapper[4696]: I1125 10:56:26.841856 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.378550 4696 generic.go:334] "Generic (PLEG): container finished" podID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerID="87d42e215c695a26e3b7b4ca984c8c2baee37d35456521565a5e09c21e03f574" exitCode=0 Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.378593 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b978c4f56-2cdxx" event={"ID":"88ab5f30-bc12-49da-8078-32be5ae2d5d9","Type":"ContainerDied","Data":"87d42e215c695a26e3b7b4ca984c8c2baee37d35456521565a5e09c21e03f574"} Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.383540 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerID="90dd0e2ba15b7ad04da0f23f9e8df06a2f527c6cabed6095b200234562d07b7f" exitCode=0 Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.383589 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerID="a67648e6342409a1ea3448b385e39d5c3f42598c4383f1fddbf65670b580c176" exitCode=2 Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.383598 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerID="6886bfd17a01518b9a498155f6ec3195289c3636b9a72d054ddc5f181733e425" exitCode=0 Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.383611 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerDied","Data":"90dd0e2ba15b7ad04da0f23f9e8df06a2f527c6cabed6095b200234562d07b7f"} Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.383645 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerDied","Data":"a67648e6342409a1ea3448b385e39d5c3f42598c4383f1fddbf65670b580c176"} Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.383654 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerDied","Data":"6886bfd17a01518b9a498155f6ec3195289c3636b9a72d054ddc5f181733e425"} Nov 25 10:56:27 crc kubenswrapper[4696]: I1125 10:56:27.392511 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78446968f5-84c89" event={"ID":"142ffa20-cd2a-4b6b-90ff-2770e6eed020","Type":"ContainerStarted","Data":"8395007b43852df3bc2ee0835f1c014d09da524694bcc6b7c4b7620191a7667e"} Nov 25 10:56:28 crc kubenswrapper[4696]: I1125 10:56:28.845625 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 10:56:29 crc kubenswrapper[4696]: I1125 10:56:29.425913 4696 generic.go:334] "Generic (PLEG): container finished" podID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerID="88a9999603fdc26e8442e67101cb67c0c71a56b11ac4283bd258f1aaf40ea1d6" exitCode=0 Nov 25 10:56:29 crc kubenswrapper[4696]: I1125 10:56:29.425975 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b978c4f56-2cdxx" event={"ID":"88ab5f30-bc12-49da-8078-32be5ae2d5d9","Type":"ContainerDied","Data":"88a9999603fdc26e8442e67101cb67c0c71a56b11ac4283bd258f1aaf40ea1d6"} Nov 25 10:56:29 crc kubenswrapper[4696]: I1125 10:56:29.524008 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.161:3000/\": dial tcp 10.217.0.161:3000: connect: connection refused" Nov 25 10:56:31 crc kubenswrapper[4696]: I1125 10:56:31.458219 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerID="ad986f8b09ab9b48845583ca4d819f7c809fdc72a205248b7a96988b2ad37292" exitCode=0 Nov 25 10:56:31 crc kubenswrapper[4696]: I1125 10:56:31.458278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerDied","Data":"ad986f8b09ab9b48845583ca4d819f7c809fdc72a205248b7a96988b2ad37292"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.147327 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258211 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-sg-core-conf-yaml\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258257 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-config-data\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258291 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-log-httpd\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258311 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-combined-ca-bundle\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258389 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqrjd\" (UniqueName: \"kubernetes.io/projected/b12fabdd-7484-4bcc-b3e5-e583e5742383-kube-api-access-gqrjd\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258463 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-scripts\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.258536 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-run-httpd\") pod \"b12fabdd-7484-4bcc-b3e5-e583e5742383\" (UID: \"b12fabdd-7484-4bcc-b3e5-e583e5742383\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.260106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.262755 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.273714 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b12fabdd-7484-4bcc-b3e5-e583e5742383-kube-api-access-gqrjd" (OuterVolumeSpecName: "kube-api-access-gqrjd") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "kube-api-access-gqrjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.301918 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-scripts" (OuterVolumeSpecName: "scripts") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.326837 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.360396 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.360418 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.360427 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b12fabdd-7484-4bcc-b3e5-e583e5742383-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.360441 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqrjd\" (UniqueName: \"kubernetes.io/projected/b12fabdd-7484-4bcc-b3e5-e583e5742383-kube-api-access-gqrjd\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.370924 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.465159 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-ovndb-tls-certs\") pod \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.465610 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-combined-ca-bundle\") pod \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.465647 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-config\") pod \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.467345 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zz6m\" (UniqueName: \"kubernetes.io/projected/88ab5f30-bc12-49da-8078-32be5ae2d5d9-kube-api-access-7zz6m\") pod \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.467384 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-httpd-config\") pod \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\" (UID: \"88ab5f30-bc12-49da-8078-32be5ae2d5d9\") " Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.467930 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.468243 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.468278 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.475273 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-config-data" (OuterVolumeSpecName: "config-data") pod "b12fabdd-7484-4bcc-b3e5-e583e5742383" (UID: "b12fabdd-7484-4bcc-b3e5-e583e5742383"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.490118 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ab5f30-bc12-49da-8078-32be5ae2d5d9-kube-api-access-7zz6m" (OuterVolumeSpecName: "kube-api-access-7zz6m") pod "88ab5f30-bc12-49da-8078-32be5ae2d5d9" (UID: "88ab5f30-bc12-49da-8078-32be5ae2d5d9"). InnerVolumeSpecName "kube-api-access-7zz6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.491871 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "88ab5f30-bc12-49da-8078-32be5ae2d5d9" (UID: "88ab5f30-bc12-49da-8078-32be5ae2d5d9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.499562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b978c4f56-2cdxx" event={"ID":"88ab5f30-bc12-49da-8078-32be5ae2d5d9","Type":"ContainerDied","Data":"8a6a4fe123b1c778b9a903312ab2257947268ebcdee24714d1c006540ffb12c8"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.499607 4696 scope.go:117] "RemoveContainer" containerID="87d42e215c695a26e3b7b4ca984c8c2baee37d35456521565a5e09c21e03f574" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.499727 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b978c4f56-2cdxx" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.506994 4696 generic.go:334] "Generic (PLEG): container finished" podID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerID="8ac4f70641365c40d7d066cd108525fd104f0eaf62202401f21b85c5a5995efc" exitCode=137 Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.507061 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6558fc797b-2qp4r" event={"ID":"1944527b-6ffc-4b6b-954f-4b01394ea0cf","Type":"ContainerDied","Data":"8ac4f70641365c40d7d066cd108525fd104f0eaf62202401f21b85c5a5995efc"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.507088 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6558fc797b-2qp4r" event={"ID":"1944527b-6ffc-4b6b-954f-4b01394ea0cf","Type":"ContainerStarted","Data":"6ecead1183cf69a63e5f49380fbd8e218673c1ab4b1f292a7094cc4436246d44"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.508806 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"14829ad2-5154-4bc3-ac0d-8343c93c0f6d","Type":"ContainerStarted","Data":"638b2795cee95202acfad0a56e5303a739ace9e27ddb4bc37e17a2e4347aa123"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.510738 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b12fabdd-7484-4bcc-b3e5-e583e5742383","Type":"ContainerDied","Data":"be0a4297ad834cc8271358c1c78b5465a3aeb15a0f9c6660ea911e10b43f8973"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.510929 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.522614 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78446968f5-84c89" event={"ID":"142ffa20-cd2a-4b6b-90ff-2770e6eed020","Type":"ContainerStarted","Data":"f755cc768fcc3482b00a94b3e26c5f3eb07def0b90047078bb2f57ad4424c233"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.523729 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.523987 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.528969 4696 scope.go:117] "RemoveContainer" containerID="88a9999603fdc26e8442e67101cb67c0c71a56b11ac4283bd258f1aaf40ea1d6" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.529516 4696 generic.go:334] "Generic (PLEG): container finished" podID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerID="b743c34bdb2c734c6424f22dad595eb14070e34899fbea0e6b459832b98e3148" exitCode=137 Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.529556 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerDied","Data":"b743c34bdb2c734c6424f22dad595eb14070e34899fbea0e6b459832b98e3148"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.529594 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerStarted","Data":"91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829"} Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.540756 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-78446968f5-84c89" podUID="142ffa20-cd2a-4b6b-90ff-2770e6eed020" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.568159 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-config" (OuterVolumeSpecName: "config") pod "88ab5f30-bc12-49da-8078-32be5ae2d5d9" (UID: "88ab5f30-bc12-49da-8078-32be5ae2d5d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.573031 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zz6m\" (UniqueName: \"kubernetes.io/projected/88ab5f30-bc12-49da-8078-32be5ae2d5d9-kube-api-access-7zz6m\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.573158 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.573215 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12fabdd-7484-4bcc-b3e5-e583e5742383-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.573269 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.577026 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88ab5f30-bc12-49da-8078-32be5ae2d5d9" (UID: "88ab5f30-bc12-49da-8078-32be5ae2d5d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.577363 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.603174583 podStartE2EDuration="14.577345681s" podCreationTimestamp="2025-11-25 10:56:19 +0000 UTC" firstStartedPulling="2025-11-25 10:56:20.932746349 +0000 UTC m=+1177.746362936" lastFinishedPulling="2025-11-25 10:56:32.906917447 +0000 UTC m=+1189.720534034" observedRunningTime="2025-11-25 10:56:33.560573665 +0000 UTC m=+1190.374190252" watchObservedRunningTime="2025-11-25 10:56:33.577345681 +0000 UTC m=+1190.390962258" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.587898 4696 scope.go:117] "RemoveContainer" containerID="90dd0e2ba15b7ad04da0f23f9e8df06a2f527c6cabed6095b200234562d07b7f" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.613819 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "88ab5f30-bc12-49da-8078-32be5ae2d5d9" (UID: "88ab5f30-bc12-49da-8078-32be5ae2d5d9"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.621794 4696 scope.go:117] "RemoveContainer" containerID="a67648e6342409a1ea3448b385e39d5c3f42598c4383f1fddbf65670b580c176" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.657939 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.662470 4696 scope.go:117] "RemoveContainer" containerID="ad986f8b09ab9b48845583ca4d819f7c809fdc72a205248b7a96988b2ad37292" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.672443 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-78446968f5-84c89" podStartSLOduration=8.672425248 podStartE2EDuration="8.672425248s" podCreationTimestamp="2025-11-25 10:56:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:33.598197497 +0000 UTC m=+1190.411814084" watchObservedRunningTime="2025-11-25 10:56:33.672425248 +0000 UTC m=+1190.486041835" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.672706 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.683752 4696 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.683793 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ab5f30-bc12-49da-8078-32be5ae2d5d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.696773 4696 scope.go:117] "RemoveContainer" containerID="6886bfd17a01518b9a498155f6ec3195289c3636b9a72d054ddc5f181733e425" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.709702 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:33 crc kubenswrapper[4696]: E1125 10:56:33.710047 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="sg-core" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710063 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="sg-core" Nov 25 10:56:33 crc kubenswrapper[4696]: E1125 10:56:33.710077 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-notification-agent" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710084 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-notification-agent" Nov 25 10:56:33 crc kubenswrapper[4696]: E1125 10:56:33.710098 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="proxy-httpd" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710104 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="proxy-httpd" Nov 25 10:56:33 crc kubenswrapper[4696]: E1125 10:56:33.710115 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-httpd" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710121 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-httpd" Nov 25 10:56:33 crc kubenswrapper[4696]: E1125 10:56:33.710142 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-central-agent" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710148 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-central-agent" Nov 25 10:56:33 crc kubenswrapper[4696]: E1125 10:56:33.710161 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-api" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710166 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-api" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710324 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-httpd" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710340 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="proxy-httpd" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710355 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-central-agent" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710363 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="sg-core" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710372 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" containerName="ceilometer-notification-agent" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.710384 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" containerName="neutron-api" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.711955 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.714463 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.717121 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.739969 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.872575 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b978c4f56-2cdxx"] Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.882875 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b978c4f56-2cdxx"] Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.887412 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-log-httpd\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.887467 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-run-httpd\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.887512 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-config-data\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.887773 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9ld\" (UniqueName: \"kubernetes.io/projected/00d1d613-20be-4911-a057-cf8e0839ac7c-kube-api-access-7q9ld\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.887905 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.888019 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.888063 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-scripts\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989459 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-log-httpd\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989514 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-run-httpd\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989556 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-config-data\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989625 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9ld\" (UniqueName: \"kubernetes.io/projected/00d1d613-20be-4911-a057-cf8e0839ac7c-kube-api-access-7q9ld\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989682 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989708 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.989724 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-scripts\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.990083 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-log-httpd\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.990644 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-run-httpd\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:33 crc kubenswrapper[4696]: I1125 10:56:33.994487 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-config-data\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:33.995325 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-scripts\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:33.998775 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.006708 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9ld\" (UniqueName: \"kubernetes.io/projected/00d1d613-20be-4911-a057-cf8e0839ac7c-kube-api-access-7q9ld\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.006878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " pod="openstack/ceilometer-0" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.030993 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.060068 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ab5f30-bc12-49da-8078-32be5ae2d5d9" path="/var/lib/kubelet/pods/88ab5f30-bc12-49da-8078-32be5ae2d5d9/volumes" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.060732 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b12fabdd-7484-4bcc-b3e5-e583e5742383" path="/var/lib/kubelet/pods/b12fabdd-7484-4bcc-b3e5-e583e5742383/volumes" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.545586 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:34 crc kubenswrapper[4696]: W1125 10:56:34.560464 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00d1d613_20be_4911_a057_cf8e0839ac7c.slice/crio-4813b8ce58d811cac3621034dabe03c827ea513f609c187c00aa09c3b81ca576 WatchSource:0}: Error finding container 4813b8ce58d811cac3621034dabe03c827ea513f609c187c00aa09c3b81ca576: Status 404 returned error can't find the container with id 4813b8ce58d811cac3621034dabe03c827ea513f609c187c00aa09c3b81ca576 Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.570060 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-78446968f5-84c89" podUID="142ffa20-cd2a-4b6b-90ff-2770e6eed020" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 10:56:34 crc kubenswrapper[4696]: I1125 10:56:34.682443 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-78446968f5-84c89" podUID="142ffa20-cd2a-4b6b-90ff-2770e6eed020" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 10:56:35 crc kubenswrapper[4696]: I1125 10:56:35.570912 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerStarted","Data":"0cfd1637bfaafca60876074a2a78cb36ba9ac2266095579a899a8bdf6fd0d578"} Nov 25 10:56:35 crc kubenswrapper[4696]: I1125 10:56:35.571537 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerStarted","Data":"4813b8ce58d811cac3621034dabe03c827ea513f609c187c00aa09c3b81ca576"} Nov 25 10:56:35 crc kubenswrapper[4696]: I1125 10:56:35.588134 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:35 crc kubenswrapper[4696]: I1125 10:56:35.752305 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.621936 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerStarted","Data":"9c8a733f726373d227026997be1af1ed9765b48943240e30736ce9a749fbb82f"} Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.645300 4696 generic.go:334] "Generic (PLEG): container finished" podID="4c1ef208-c90a-458e-a128-e89476c69a33" containerID="f701dbc71200da9e920f1193f1ded1c9dc3ffd72ba0a90f6105be22e474e00aa" exitCode=137 Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.646961 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4c1ef208-c90a-458e-a128-e89476c69a33","Type":"ContainerDied","Data":"f701dbc71200da9e920f1193f1ded1c9dc3ffd72ba0a90f6105be22e474e00aa"} Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.679215 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.848312 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r57s4\" (UniqueName: \"kubernetes.io/projected/4c1ef208-c90a-458e-a128-e89476c69a33-kube-api-access-r57s4\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.849319 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-combined-ca-bundle\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.849366 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.849402 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c1ef208-c90a-458e-a128-e89476c69a33-etc-machine-id\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.849465 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data-custom\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.849559 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c1ef208-c90a-458e-a128-e89476c69a33-logs\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.849598 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-scripts\") pod \"4c1ef208-c90a-458e-a128-e89476c69a33\" (UID: \"4c1ef208-c90a-458e-a128-e89476c69a33\") " Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.851767 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c1ef208-c90a-458e-a128-e89476c69a33-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.852902 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c1ef208-c90a-458e-a128-e89476c69a33-kube-api-access-r57s4" (OuterVolumeSpecName: "kube-api-access-r57s4") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "kube-api-access-r57s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.853240 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c1ef208-c90a-458e-a128-e89476c69a33-logs" (OuterVolumeSpecName: "logs") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.863785 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-scripts" (OuterVolumeSpecName: "scripts") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.885105 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.925155 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.952073 4696 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c1ef208-c90a-458e-a128-e89476c69a33-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.952104 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.952115 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c1ef208-c90a-458e-a128-e89476c69a33-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.952125 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.952148 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r57s4\" (UniqueName: \"kubernetes.io/projected/4c1ef208-c90a-458e-a128-e89476c69a33-kube-api-access-r57s4\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.952163 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:36 crc kubenswrapper[4696]: I1125 10:56:36.962015 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data" (OuterVolumeSpecName: "config-data") pod "4c1ef208-c90a-458e-a128-e89476c69a33" (UID: "4c1ef208-c90a-458e-a128-e89476c69a33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.053456 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1ef208-c90a-458e-a128-e89476c69a33-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.657089 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4c1ef208-c90a-458e-a128-e89476c69a33","Type":"ContainerDied","Data":"13624b60027b04bfb610e4780064580abbfb8cf8091eb4156c281daf441964b1"} Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.657418 4696 scope.go:117] "RemoveContainer" containerID="f701dbc71200da9e920f1193f1ded1c9dc3ffd72ba0a90f6105be22e474e00aa" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.657216 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.661093 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerStarted","Data":"2a094b9ab0ebf204057d762e4bd52c0f4ccdeda4700aef0f5ed41d91d214012f"} Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.686765 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.696987 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.715348 4696 scope.go:117] "RemoveContainer" containerID="1f854fc5e2148b76839c7980b26081f0bdf9f9141e2bec1987935c777037ba23" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.717557 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:37 crc kubenswrapper[4696]: E1125 10:56:37.717929 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.717946 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api" Nov 25 10:56:37 crc kubenswrapper[4696]: E1125 10:56:37.717982 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api-log" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.717989 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api-log" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.718163 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.718199 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" containerName="cinder-api-log" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.719340 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.726116 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.726364 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.726975 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.733537 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.867877 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-scripts\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.867929 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.867962 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.867998 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-logs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.868057 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-config-data\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.868105 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.868137 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-config-data-custom\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.868301 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.868382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlztg\" (UniqueName: \"kubernetes.io/projected/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-kube-api-access-zlztg\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.970204 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-config-data-custom\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.970264 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.970293 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlztg\" (UniqueName: \"kubernetes.io/projected/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-kube-api-access-zlztg\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971029 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-scripts\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971075 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971100 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971141 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-logs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971186 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-config-data\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971296 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.971884 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-logs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.980288 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-config-data-custom\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.982012 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.986639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-config-data\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.987401 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.988127 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.991732 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-scripts\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:37 crc kubenswrapper[4696]: I1125 10:56:37.994063 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlztg\" (UniqueName: \"kubernetes.io/projected/5a8e77cd-7f3d-45e6-bb36-2ce208f70adb-kube-api-access-zlztg\") pod \"cinder-api-0\" (UID: \"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb\") " pod="openstack/cinder-api-0" Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.047832 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.051731 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c1ef208-c90a-458e-a128-e89476c69a33" path="/var/lib/kubelet/pods/4c1ef208-c90a-458e-a128-e89476c69a33/volumes" Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.371439 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.693983 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerStarted","Data":"63eb6298808ebed022dc701c8c43125972bf256e232142ab4b08c98aa759903f"} Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.694487 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-central-agent" containerID="cri-o://0cfd1637bfaafca60876074a2a78cb36ba9ac2266095579a899a8bdf6fd0d578" gracePeriod=30 Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.694577 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.694981 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="proxy-httpd" containerID="cri-o://63eb6298808ebed022dc701c8c43125972bf256e232142ab4b08c98aa759903f" gracePeriod=30 Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.695036 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="sg-core" containerID="cri-o://2a094b9ab0ebf204057d762e4bd52c0f4ccdeda4700aef0f5ed41d91d214012f" gracePeriod=30 Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.695086 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-notification-agent" containerID="cri-o://9c8a733f726373d227026997be1af1ed9765b48943240e30736ce9a749fbb82f" gracePeriod=30 Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.705134 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb","Type":"ContainerStarted","Data":"59801e49f56c6364d889101abfedf024bcbc5dba71fcb6dfab08f844cf559d3b"} Nov 25 10:56:38 crc kubenswrapper[4696]: I1125 10:56:38.735427 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.09698889 podStartE2EDuration="5.735405335s" podCreationTimestamp="2025-11-25 10:56:33 +0000 UTC" firstStartedPulling="2025-11-25 10:56:34.573013994 +0000 UTC m=+1191.386630581" lastFinishedPulling="2025-11-25 10:56:38.211430439 +0000 UTC m=+1195.025047026" observedRunningTime="2025-11-25 10:56:38.73118631 +0000 UTC m=+1195.544802897" watchObservedRunningTime="2025-11-25 10:56:38.735405335 +0000 UTC m=+1195.549021922" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.369379 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hdlx8"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.372756 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.385003 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hdlx8"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.490288 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-x9l8x"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.492267 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.517089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ba9c13d-1506-46fd-822e-4d816ecff490-operator-scripts\") pod \"nova-api-db-create-hdlx8\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.517345 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98f8l\" (UniqueName: \"kubernetes.io/projected/2ba9c13d-1506-46fd-822e-4d816ecff490-kube-api-access-98f8l\") pod \"nova-api-db-create-hdlx8\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.565612 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x9l8x"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.630374 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ba9c13d-1506-46fd-822e-4d816ecff490-operator-scripts\") pod \"nova-api-db-create-hdlx8\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.630470 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckx42\" (UniqueName: \"kubernetes.io/projected/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-kube-api-access-ckx42\") pod \"nova-cell0-db-create-x9l8x\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.630530 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98f8l\" (UniqueName: \"kubernetes.io/projected/2ba9c13d-1506-46fd-822e-4d816ecff490-kube-api-access-98f8l\") pod \"nova-api-db-create-hdlx8\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.630592 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-operator-scripts\") pod \"nova-cell0-db-create-x9l8x\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.632195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ba9c13d-1506-46fd-822e-4d816ecff490-operator-scripts\") pod \"nova-api-db-create-hdlx8\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.653169 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-2731-account-create-tprt5"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.658709 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.661975 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.672466 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2731-account-create-tprt5"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.687835 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98f8l\" (UniqueName: \"kubernetes.io/projected/2ba9c13d-1506-46fd-822e-4d816ecff490-kube-api-access-98f8l\") pod \"nova-api-db-create-hdlx8\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.715811 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-plnjv"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.718420 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.732875 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-operator-scripts\") pod \"nova-cell0-db-create-x9l8x\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.733304 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckx42\" (UniqueName: \"kubernetes.io/projected/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-kube-api-access-ckx42\") pod \"nova-cell0-db-create-x9l8x\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.735820 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-plnjv"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.738206 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-operator-scripts\") pod \"nova-cell0-db-create-x9l8x\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.740661 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.766659 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckx42\" (UniqueName: \"kubernetes.io/projected/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-kube-api-access-ckx42\") pod \"nova-cell0-db-create-x9l8x\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.769101 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-162c-account-create-sn7k7"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.769991 4696 generic.go:334] "Generic (PLEG): container finished" podID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerID="63eb6298808ebed022dc701c8c43125972bf256e232142ab4b08c98aa759903f" exitCode=0 Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.770026 4696 generic.go:334] "Generic (PLEG): container finished" podID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerID="2a094b9ab0ebf204057d762e4bd52c0f4ccdeda4700aef0f5ed41d91d214012f" exitCode=2 Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.770039 4696 generic.go:334] "Generic (PLEG): container finished" podID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerID="9c8a733f726373d227026997be1af1ed9765b48943240e30736ce9a749fbb82f" exitCode=0 Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.771272 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerDied","Data":"63eb6298808ebed022dc701c8c43125972bf256e232142ab4b08c98aa759903f"} Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.771317 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerDied","Data":"2a094b9ab0ebf204057d762e4bd52c0f4ccdeda4700aef0f5ed41d91d214012f"} Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.771329 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerDied","Data":"9c8a733f726373d227026997be1af1ed9765b48943240e30736ce9a749fbb82f"} Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.771407 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.773381 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.801374 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-162c-account-create-sn7k7"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.802297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb","Type":"ContainerStarted","Data":"80729348b606a02ff1d66d74a113b2e81d85888f0c1b4266466ce5b61ea666c3"} Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.834987 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-operator-scripts\") pod \"nova-api-2731-account-create-tprt5\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.835038 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-operator-scripts\") pod \"nova-cell1-db-create-plnjv\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.835152 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwzmq\" (UniqueName: \"kubernetes.io/projected/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-kube-api-access-fwzmq\") pod \"nova-cell1-db-create-plnjv\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.835208 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hfpm\" (UniqueName: \"kubernetes.io/projected/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-kube-api-access-4hfpm\") pod \"nova-api-2731-account-create-tprt5\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.934026 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ac60-account-create-grcfr"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.935469 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.936260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwzmq\" (UniqueName: \"kubernetes.io/projected/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-kube-api-access-fwzmq\") pod \"nova-cell1-db-create-plnjv\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.936296 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5ql6\" (UniqueName: \"kubernetes.io/projected/b183e265-9749-45f5-9b87-de239a7daca0-kube-api-access-l5ql6\") pod \"nova-cell0-162c-account-create-sn7k7\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.936348 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hfpm\" (UniqueName: \"kubernetes.io/projected/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-kube-api-access-4hfpm\") pod \"nova-api-2731-account-create-tprt5\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.936382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b183e265-9749-45f5-9b87-de239a7daca0-operator-scripts\") pod \"nova-cell0-162c-account-create-sn7k7\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.936405 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-operator-scripts\") pod \"nova-api-2731-account-create-tprt5\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.936428 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-operator-scripts\") pod \"nova-cell1-db-create-plnjv\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.937102 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-operator-scripts\") pod \"nova-cell1-db-create-plnjv\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.937198 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-operator-scripts\") pod \"nova-api-2731-account-create-tprt5\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.942969 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.962503 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.970034 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwzmq\" (UniqueName: \"kubernetes.io/projected/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-kube-api-access-fwzmq\") pod \"nova-cell1-db-create-plnjv\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.983230 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ac60-account-create-grcfr"] Nov 25 10:56:39 crc kubenswrapper[4696]: I1125 10:56:39.984917 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hfpm\" (UniqueName: \"kubernetes.io/projected/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-kube-api-access-4hfpm\") pod \"nova-api-2731-account-create-tprt5\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.016285 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.041297 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5ql6\" (UniqueName: \"kubernetes.io/projected/b183e265-9749-45f5-9b87-de239a7daca0-kube-api-access-l5ql6\") pod \"nova-cell0-162c-account-create-sn7k7\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.041395 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b183e265-9749-45f5-9b87-de239a7daca0-operator-scripts\") pod \"nova-cell0-162c-account-create-sn7k7\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.041428 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b276d474-ccfa-4b14-95e6-a92d68897387-operator-scripts\") pod \"nova-cell1-ac60-account-create-grcfr\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.041487 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrnl9\" (UniqueName: \"kubernetes.io/projected/b276d474-ccfa-4b14-95e6-a92d68897387-kube-api-access-vrnl9\") pod \"nova-cell1-ac60-account-create-grcfr\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.044485 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b183e265-9749-45f5-9b87-de239a7daca0-operator-scripts\") pod \"nova-cell0-162c-account-create-sn7k7\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.064092 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.072651 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5ql6\" (UniqueName: \"kubernetes.io/projected/b183e265-9749-45f5-9b87-de239a7daca0-kube-api-access-l5ql6\") pod \"nova-cell0-162c-account-create-sn7k7\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.090380 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.143910 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b276d474-ccfa-4b14-95e6-a92d68897387-operator-scripts\") pod \"nova-cell1-ac60-account-create-grcfr\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.144006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrnl9\" (UniqueName: \"kubernetes.io/projected/b276d474-ccfa-4b14-95e6-a92d68897387-kube-api-access-vrnl9\") pod \"nova-cell1-ac60-account-create-grcfr\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.144654 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b276d474-ccfa-4b14-95e6-a92d68897387-operator-scripts\") pod \"nova-cell1-ac60-account-create-grcfr\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.166015 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrnl9\" (UniqueName: \"kubernetes.io/projected/b276d474-ccfa-4b14-95e6-a92d68897387-kube-api-access-vrnl9\") pod \"nova-cell1-ac60-account-create-grcfr\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.281149 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.459746 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hdlx8"] Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.684828 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78446968f5-84c89" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.791966 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-plnjv"] Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.818436 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2731-account-create-tprt5"] Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.823209 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5a8e77cd-7f3d-45e6-bb36-2ce208f70adb","Type":"ContainerStarted","Data":"9140ee25a872e320e3aed1ed95753bf7dda8830ba067f186eb694d0fec79a6fc"} Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.823455 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.841503 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdlx8" event={"ID":"2ba9c13d-1506-46fd-822e-4d816ecff490","Type":"ContainerStarted","Data":"dcb0238dd2109f37a083c8a5e09012b1ee582d563e8545ca7fc0a57b00729346"} Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.841580 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdlx8" event={"ID":"2ba9c13d-1506-46fd-822e-4d816ecff490","Type":"ContainerStarted","Data":"593f4659c6c9061a3efdf1d033d1376b1e11541340b705ff112cadb798fe7e32"} Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.862820 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.8628030349999998 podStartE2EDuration="3.862803035s" podCreationTimestamp="2025-11-25 10:56:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:40.857751304 +0000 UTC m=+1197.671367911" watchObservedRunningTime="2025-11-25 10:56:40.862803035 +0000 UTC m=+1197.676419622" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.864729 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnjv" event={"ID":"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0","Type":"ContainerStarted","Data":"d4ec6f454ef98c4a09cfbcd96361fbe9e7027ccdfc7879104a54029498315e39"} Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.891823 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-162c-account-create-sn7k7"] Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.908148 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-hdlx8" podStartSLOduration=1.908124733 podStartE2EDuration="1.908124733s" podCreationTimestamp="2025-11-25 10:56:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:40.877599648 +0000 UTC m=+1197.691216225" watchObservedRunningTime="2025-11-25 10:56:40.908124733 +0000 UTC m=+1197.721741320" Nov 25 10:56:40 crc kubenswrapper[4696]: I1125 10:56:40.927709 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x9l8x"] Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.117721 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ac60-account-create-grcfr"] Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.879729 4696 generic.go:334] "Generic (PLEG): container finished" podID="b183e265-9749-45f5-9b87-de239a7daca0" containerID="541f33e91f124886097bff5992fb8bbec7619b894e803294c7fd06e12c9c5e6f" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.880052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-162c-account-create-sn7k7" event={"ID":"b183e265-9749-45f5-9b87-de239a7daca0","Type":"ContainerDied","Data":"541f33e91f124886097bff5992fb8bbec7619b894e803294c7fd06e12c9c5e6f"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.880087 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-162c-account-create-sn7k7" event={"ID":"b183e265-9749-45f5-9b87-de239a7daca0","Type":"ContainerStarted","Data":"a60f05994848b96ad91a61a14565af0537bdb794f25870d5c828398ccada3d61"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.883775 4696 generic.go:334] "Generic (PLEG): container finished" podID="bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" containerID="49b5f5810bc0b4b182f8276f34fca35c5498ba49f506a497888c582525a02b21" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.883825 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnjv" event={"ID":"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0","Type":"ContainerDied","Data":"49b5f5810bc0b4b182f8276f34fca35c5498ba49f506a497888c582525a02b21"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.885451 4696 generic.go:334] "Generic (PLEG): container finished" podID="95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" containerID="704497debdd2b3d4266c02940295a55717e3f4ded974697b047eccb74c8eeb1e" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.885502 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x9l8x" event={"ID":"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19","Type":"ContainerDied","Data":"704497debdd2b3d4266c02940295a55717e3f4ded974697b047eccb74c8eeb1e"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.885520 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x9l8x" event={"ID":"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19","Type":"ContainerStarted","Data":"a466045312fc201fc013bef32d189833b744d2908fc242bdaed9d3ec34d86570"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.887131 4696 generic.go:334] "Generic (PLEG): container finished" podID="636b96ee-a180-4234-9a9b-f1ebfd9a2f25" containerID="1c344fc7b37117bd87c2080daf5416fc850b91049a7cd609bac58ee740817368" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.887174 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2731-account-create-tprt5" event={"ID":"636b96ee-a180-4234-9a9b-f1ebfd9a2f25","Type":"ContainerDied","Data":"1c344fc7b37117bd87c2080daf5416fc850b91049a7cd609bac58ee740817368"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.887190 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2731-account-create-tprt5" event={"ID":"636b96ee-a180-4234-9a9b-f1ebfd9a2f25","Type":"ContainerStarted","Data":"4164a07aaedf728762939f67933e8f8d4a6ec5b8131135da07844e0fa6116c5d"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.895210 4696 generic.go:334] "Generic (PLEG): container finished" podID="b276d474-ccfa-4b14-95e6-a92d68897387" containerID="b067cca4b293436a9051abdb47d1fa1c8ce3689f242443c36f855fb3b5b27d3d" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.895295 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac60-account-create-grcfr" event={"ID":"b276d474-ccfa-4b14-95e6-a92d68897387","Type":"ContainerDied","Data":"b067cca4b293436a9051abdb47d1fa1c8ce3689f242443c36f855fb3b5b27d3d"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.895319 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac60-account-create-grcfr" event={"ID":"b276d474-ccfa-4b14-95e6-a92d68897387","Type":"ContainerStarted","Data":"6afe10df165237bb8c9bc78a485e3b3b05cd0c0b132afaa7418352316576e23f"} Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.901354 4696 generic.go:334] "Generic (PLEG): container finished" podID="2ba9c13d-1506-46fd-822e-4d816ecff490" containerID="dcb0238dd2109f37a083c8a5e09012b1ee582d563e8545ca7fc0a57b00729346" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4696]: I1125 10:56:41.901675 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdlx8" event={"ID":"2ba9c13d-1506-46fd-822e-4d816ecff490","Type":"ContainerDied","Data":"dcb0238dd2109f37a083c8a5e09012b1ee582d563e8545ca7fc0a57b00729346"} Nov 25 10:56:42 crc kubenswrapper[4696]: I1125 10:56:42.760694 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:56:42 crc kubenswrapper[4696]: I1125 10:56:42.760749 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:56:42 crc kubenswrapper[4696]: I1125 10:56:42.876420 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:56:42 crc kubenswrapper[4696]: I1125 10:56:42.876469 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.419010 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.501645 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.502285 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-log" containerID="cri-o://5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40" gracePeriod=30 Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.502476 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-httpd" containerID="cri-o://dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff" gracePeriod=30 Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.529146 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckx42\" (UniqueName: \"kubernetes.io/projected/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-kube-api-access-ckx42\") pod \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.529353 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-operator-scripts\") pod \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\" (UID: \"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.530361 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" (UID: "95c4ea9f-5c8c-4475-9c5c-3c00ef595f19"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.591829 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-kube-api-access-ckx42" (OuterVolumeSpecName: "kube-api-access-ckx42") pod "95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" (UID: "95c4ea9f-5c8c-4475-9c5c-3c00ef595f19"). InnerVolumeSpecName "kube-api-access-ckx42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.655781 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckx42\" (UniqueName: \"kubernetes.io/projected/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-kube-api-access-ckx42\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.655823 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.777545 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.809975 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.814641 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.830163 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.859482 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.859801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hfpm\" (UniqueName: \"kubernetes.io/projected/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-kube-api-access-4hfpm\") pod \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.859935 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-operator-scripts\") pod \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.859986 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwzmq\" (UniqueName: \"kubernetes.io/projected/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-kube-api-access-fwzmq\") pod \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\" (UID: \"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.860086 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-operator-scripts\") pod \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\" (UID: \"636b96ee-a180-4234-9a9b-f1ebfd9a2f25\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.861963 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" (UID: "bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.863534 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "636b96ee-a180-4234-9a9b-f1ebfd9a2f25" (UID: "636b96ee-a180-4234-9a9b-f1ebfd9a2f25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.868609 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-kube-api-access-fwzmq" (OuterVolumeSpecName: "kube-api-access-fwzmq") pod "bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" (UID: "bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0"). InnerVolumeSpecName "kube-api-access-fwzmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.868476 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-kube-api-access-4hfpm" (OuterVolumeSpecName: "kube-api-access-4hfpm") pod "636b96ee-a180-4234-9a9b-f1ebfd9a2f25" (UID: "636b96ee-a180-4234-9a9b-f1ebfd9a2f25"). InnerVolumeSpecName "kube-api-access-4hfpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.924169 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnjv" event={"ID":"bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0","Type":"ContainerDied","Data":"d4ec6f454ef98c4a09cfbcd96361fbe9e7027ccdfc7879104a54029498315e39"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.924206 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4ec6f454ef98c4a09cfbcd96361fbe9e7027ccdfc7879104a54029498315e39" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.924254 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnjv" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.928323 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x9l8x" event={"ID":"95c4ea9f-5c8c-4475-9c5c-3c00ef595f19","Type":"ContainerDied","Data":"a466045312fc201fc013bef32d189833b744d2908fc242bdaed9d3ec34d86570"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.928362 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a466045312fc201fc013bef32d189833b744d2908fc242bdaed9d3ec34d86570" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.928413 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x9l8x" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.934324 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2731-account-create-tprt5" event={"ID":"636b96ee-a180-4234-9a9b-f1ebfd9a2f25","Type":"ContainerDied","Data":"4164a07aaedf728762939f67933e8f8d4a6ec5b8131135da07844e0fa6116c5d"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.934355 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4164a07aaedf728762939f67933e8f8d4a6ec5b8131135da07844e0fa6116c5d" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.934398 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2731-account-create-tprt5" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.941802 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac60-account-create-grcfr" event={"ID":"b276d474-ccfa-4b14-95e6-a92d68897387","Type":"ContainerDied","Data":"6afe10df165237bb8c9bc78a485e3b3b05cd0c0b132afaa7418352316576e23f"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.941840 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6afe10df165237bb8c9bc78a485e3b3b05cd0c0b132afaa7418352316576e23f" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.941897 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac60-account-create-grcfr" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.949033 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdlx8" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.949417 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdlx8" event={"ID":"2ba9c13d-1506-46fd-822e-4d816ecff490","Type":"ContainerDied","Data":"593f4659c6c9061a3efdf1d033d1376b1e11541340b705ff112cadb798fe7e32"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.949445 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="593f4659c6c9061a3efdf1d033d1376b1e11541340b705ff112cadb798fe7e32" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.951152 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-162c-account-create-sn7k7" event={"ID":"b183e265-9749-45f5-9b87-de239a7daca0","Type":"ContainerDied","Data":"a60f05994848b96ad91a61a14565af0537bdb794f25870d5c828398ccada3d61"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.951191 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a60f05994848b96ad91a61a14565af0537bdb794f25870d5c828398ccada3d61" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.951264 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-162c-account-create-sn7k7" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.955851 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerID="5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40" exitCode=143 Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.955914 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42","Type":"ContainerDied","Data":"5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40"} Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962004 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b276d474-ccfa-4b14-95e6-a92d68897387-operator-scripts\") pod \"b276d474-ccfa-4b14-95e6-a92d68897387\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962151 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5ql6\" (UniqueName: \"kubernetes.io/projected/b183e265-9749-45f5-9b87-de239a7daca0-kube-api-access-l5ql6\") pod \"b183e265-9749-45f5-9b87-de239a7daca0\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962201 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ba9c13d-1506-46fd-822e-4d816ecff490-operator-scripts\") pod \"2ba9c13d-1506-46fd-822e-4d816ecff490\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962222 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98f8l\" (UniqueName: \"kubernetes.io/projected/2ba9c13d-1506-46fd-822e-4d816ecff490-kube-api-access-98f8l\") pod \"2ba9c13d-1506-46fd-822e-4d816ecff490\" (UID: \"2ba9c13d-1506-46fd-822e-4d816ecff490\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962322 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b183e265-9749-45f5-9b87-de239a7daca0-operator-scripts\") pod \"b183e265-9749-45f5-9b87-de239a7daca0\" (UID: \"b183e265-9749-45f5-9b87-de239a7daca0\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962380 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrnl9\" (UniqueName: \"kubernetes.io/projected/b276d474-ccfa-4b14-95e6-a92d68897387-kube-api-access-vrnl9\") pod \"b276d474-ccfa-4b14-95e6-a92d68897387\" (UID: \"b276d474-ccfa-4b14-95e6-a92d68897387\") " Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962750 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962766 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwzmq\" (UniqueName: \"kubernetes.io/projected/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0-kube-api-access-fwzmq\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962807 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.962816 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hfpm\" (UniqueName: \"kubernetes.io/projected/636b96ee-a180-4234-9a9b-f1ebfd9a2f25-kube-api-access-4hfpm\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.964018 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b183e265-9749-45f5-9b87-de239a7daca0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b183e265-9749-45f5-9b87-de239a7daca0" (UID: "b183e265-9749-45f5-9b87-de239a7daca0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.964422 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b276d474-ccfa-4b14-95e6-a92d68897387-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b276d474-ccfa-4b14-95e6-a92d68897387" (UID: "b276d474-ccfa-4b14-95e6-a92d68897387"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.964606 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ba9c13d-1506-46fd-822e-4d816ecff490-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2ba9c13d-1506-46fd-822e-4d816ecff490" (UID: "2ba9c13d-1506-46fd-822e-4d816ecff490"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.968161 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b183e265-9749-45f5-9b87-de239a7daca0-kube-api-access-l5ql6" (OuterVolumeSpecName: "kube-api-access-l5ql6") pod "b183e265-9749-45f5-9b87-de239a7daca0" (UID: "b183e265-9749-45f5-9b87-de239a7daca0"). InnerVolumeSpecName "kube-api-access-l5ql6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.969463 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b276d474-ccfa-4b14-95e6-a92d68897387-kube-api-access-vrnl9" (OuterVolumeSpecName: "kube-api-access-vrnl9") pod "b276d474-ccfa-4b14-95e6-a92d68897387" (UID: "b276d474-ccfa-4b14-95e6-a92d68897387"). InnerVolumeSpecName "kube-api-access-vrnl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:43 crc kubenswrapper[4696]: I1125 10:56:43.974171 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba9c13d-1506-46fd-822e-4d816ecff490-kube-api-access-98f8l" (OuterVolumeSpecName: "kube-api-access-98f8l") pod "2ba9c13d-1506-46fd-822e-4d816ecff490" (UID: "2ba9c13d-1506-46fd-822e-4d816ecff490"). InnerVolumeSpecName "kube-api-access-98f8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.065097 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5ql6\" (UniqueName: \"kubernetes.io/projected/b183e265-9749-45f5-9b87-de239a7daca0-kube-api-access-l5ql6\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.065463 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ba9c13d-1506-46fd-822e-4d816ecff490-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.065478 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98f8l\" (UniqueName: \"kubernetes.io/projected/2ba9c13d-1506-46fd-822e-4d816ecff490-kube-api-access-98f8l\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.065519 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b183e265-9749-45f5-9b87-de239a7daca0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.065533 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrnl9\" (UniqueName: \"kubernetes.io/projected/b276d474-ccfa-4b14-95e6-a92d68897387-kube-api-access-vrnl9\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.065545 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b276d474-ccfa-4b14-95e6-a92d68897387-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.837834 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.838063 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-log" containerID="cri-o://541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b" gracePeriod=30 Nov 25 10:56:44 crc kubenswrapper[4696]: I1125 10:56:44.838159 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-httpd" containerID="cri-o://84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae" gracePeriod=30 Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.002738 4696 generic.go:334] "Generic (PLEG): container finished" podID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerID="0cfd1637bfaafca60876074a2a78cb36ba9ac2266095579a899a8bdf6fd0d578" exitCode=0 Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.002829 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerDied","Data":"0cfd1637bfaafca60876074a2a78cb36ba9ac2266095579a899a8bdf6fd0d578"} Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.005245 4696 generic.go:334] "Generic (PLEG): container finished" podID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerID="541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b" exitCode=143 Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.005279 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf","Type":"ContainerDied","Data":"541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b"} Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.656734 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717336 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-run-httpd\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717471 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-log-httpd\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717511 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-config-data\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717537 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9ld\" (UniqueName: \"kubernetes.io/projected/00d1d613-20be-4911-a057-cf8e0839ac7c-kube-api-access-7q9ld\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717627 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-scripts\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717651 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-combined-ca-bundle\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.717709 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-sg-core-conf-yaml\") pod \"00d1d613-20be-4911-a057-cf8e0839ac7c\" (UID: \"00d1d613-20be-4911-a057-cf8e0839ac7c\") " Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.719186 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.719969 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.725897 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d1d613-20be-4911-a057-cf8e0839ac7c-kube-api-access-7q9ld" (OuterVolumeSpecName: "kube-api-access-7q9ld") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "kube-api-access-7q9ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.726950 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-scripts" (OuterVolumeSpecName: "scripts") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.772940 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.820121 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9ld\" (UniqueName: \"kubernetes.io/projected/00d1d613-20be-4911-a057-cf8e0839ac7c-kube-api-access-7q9ld\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.820429 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.820519 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.820601 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.820699 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00d1d613-20be-4911-a057-cf8e0839ac7c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.854247 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.896854 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-config-data" (OuterVolumeSpecName: "config-data") pod "00d1d613-20be-4911-a057-cf8e0839ac7c" (UID: "00d1d613-20be-4911-a057-cf8e0839ac7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.922412 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:46 crc kubenswrapper[4696]: I1125 10:56:46.922643 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d1d613-20be-4911-a057-cf8e0839ac7c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.016599 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00d1d613-20be-4911-a057-cf8e0839ac7c","Type":"ContainerDied","Data":"4813b8ce58d811cac3621034dabe03c827ea513f609c187c00aa09c3b81ca576"} Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.016653 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.016659 4696 scope.go:117] "RemoveContainer" containerID="63eb6298808ebed022dc701c8c43125972bf256e232142ab4b08c98aa759903f" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.043385 4696 scope.go:117] "RemoveContainer" containerID="2a094b9ab0ebf204057d762e4bd52c0f4ccdeda4700aef0f5ed41d91d214012f" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.079792 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.097024 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108095 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108224 4696 scope.go:117] "RemoveContainer" containerID="9c8a733f726373d227026997be1af1ed9765b48943240e30736ce9a749fbb82f" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108495 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b276d474-ccfa-4b14-95e6-a92d68897387" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108509 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b276d474-ccfa-4b14-95e6-a92d68897387" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108524 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-notification-agent" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108530 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-notification-agent" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108541 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b183e265-9749-45f5-9b87-de239a7daca0" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108546 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b183e265-9749-45f5-9b87-de239a7daca0" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108558 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108564 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108571 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636b96ee-a180-4234-9a9b-f1ebfd9a2f25" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108576 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="636b96ee-a180-4234-9a9b-f1ebfd9a2f25" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108597 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108603 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108612 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="sg-core" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108618 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="sg-core" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108628 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-central-agent" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108636 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-central-agent" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108651 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="proxy-httpd" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108657 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="proxy-httpd" Nov 25 10:56:47 crc kubenswrapper[4696]: E1125 10:56:47.108680 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba9c13d-1506-46fd-822e-4d816ecff490" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108686 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba9c13d-1506-46fd-822e-4d816ecff490" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108856 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b276d474-ccfa-4b14-95e6-a92d68897387" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108872 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-notification-agent" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108882 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108891 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="636b96ee-a180-4234-9a9b-f1ebfd9a2f25" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108901 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba9c13d-1506-46fd-822e-4d816ecff490" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108911 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" containerName="mariadb-database-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108922 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="sg-core" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108930 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b183e265-9749-45f5-9b87-de239a7daca0" containerName="mariadb-account-create" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108940 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="ceilometer-central-agent" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.108947 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" containerName="proxy-httpd" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.111965 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.115824 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.115978 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.135134 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.193094 4696 scope.go:117] "RemoveContainer" containerID="0cfd1637bfaafca60876074a2a78cb36ba9ac2266095579a899a8bdf6fd0d578" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.227448 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.227754 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.227928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdt5p\" (UniqueName: \"kubernetes.io/projected/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-kube-api-access-fdt5p\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.228034 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-config-data\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.228216 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-log-httpd\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.228264 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-scripts\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.230040 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-run-httpd\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331508 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-run-httpd\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331577 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331614 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331637 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdt5p\" (UniqueName: \"kubernetes.io/projected/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-kube-api-access-fdt5p\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331654 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-config-data\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331803 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-log-httpd\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.331850 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-scripts\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.337374 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-log-httpd\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.337441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-run-httpd\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.337901 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.337946 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-scripts\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.343162 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-config-data\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.344168 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.355472 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdt5p\" (UniqueName: \"kubernetes.io/projected/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-kube-api-access-fdt5p\") pod \"ceilometer-0\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.466189 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.588365 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.729784 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.843702 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-public-tls-certs\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.843869 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-combined-ca-bundle\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.843902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc6hg\" (UniqueName: \"kubernetes.io/projected/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-kube-api-access-wc6hg\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.843932 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-logs\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.844019 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-scripts\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.844113 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.844143 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-config-data\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.844173 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-httpd-run\") pod \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\" (UID: \"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42\") " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.846103 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-logs" (OuterVolumeSpecName: "logs") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.846887 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.850853 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-kube-api-access-wc6hg" (OuterVolumeSpecName: "kube-api-access-wc6hg") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "kube-api-access-wc6hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.873798 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-scripts" (OuterVolumeSpecName: "scripts") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.912082 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.946513 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc6hg\" (UniqueName: \"kubernetes.io/projected/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-kube-api-access-wc6hg\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.946545 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.946557 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.946582 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.946596 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.946903 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.956812 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-config-data" (OuterVolumeSpecName: "config-data") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.976764 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 10:56:47 crc kubenswrapper[4696]: I1125 10:56:47.984310 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" (UID: "ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.058978 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerID="dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff" exitCode=0 Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.059118 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.067489 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.067525 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.067536 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.067544 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.087716 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d1d613-20be-4911-a057-cf8e0839ac7c" path="/var/lib/kubelet/pods/00d1d613-20be-4911-a057-cf8e0839ac7c/volumes" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.105639 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42","Type":"ContainerDied","Data":"dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff"} Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.105705 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42","Type":"ContainerDied","Data":"23d176541e9c979c5e5c57cacc68d845ce30274a64abc9327dee79b78bd75930"} Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.105839 4696 scope.go:117] "RemoveContainer" containerID="dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.132109 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.140716 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.158690 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.159846 4696 scope.go:117] "RemoveContainer" containerID="5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.177447 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:56:48 crc kubenswrapper[4696]: E1125 10:56:48.177934 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-log" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.177966 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-log" Nov 25 10:56:48 crc kubenswrapper[4696]: E1125 10:56:48.177976 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-httpd" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.177983 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-httpd" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.178234 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-log" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.178253 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" containerName="glance-httpd" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.179545 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.187222 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.188211 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.188718 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.223931 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.262111 4696 scope.go:117] "RemoveContainer" containerID="dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff" Nov 25 10:56:48 crc kubenswrapper[4696]: E1125 10:56:48.262491 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff\": container with ID starting with dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff not found: ID does not exist" containerID="dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.262520 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff"} err="failed to get container status \"dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff\": rpc error: code = NotFound desc = could not find container \"dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff\": container with ID starting with dc312c8192b72d2a00a15a94e2d649af39ec3e7f78ed7be7e447fd153d144aff not found: ID does not exist" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.262541 4696 scope.go:117] "RemoveContainer" containerID="5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40" Nov 25 10:56:48 crc kubenswrapper[4696]: E1125 10:56:48.262730 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40\": container with ID starting with 5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40 not found: ID does not exist" containerID="5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.262750 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40"} err="failed to get container status \"5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40\": rpc error: code = NotFound desc = could not find container \"5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40\": container with ID starting with 5a90da7876f4801700867ffb13b599a234189c6e23f6aa25fbbbd0b576ac0d40 not found: ID does not exist" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.377730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.377870 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.377928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-scripts\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.377974 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xltwv\" (UniqueName: \"kubernetes.io/projected/d642739c-282a-4644-9aae-68f82c3ddf58-kube-api-access-xltwv\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.378085 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-config-data\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.378166 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d642739c-282a-4644-9aae-68f82c3ddf58-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.378230 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.378320 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d642739c-282a-4644-9aae-68f82c3ddf58-logs\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.479919 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.479971 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-scripts\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.479995 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xltwv\" (UniqueName: \"kubernetes.io/projected/d642739c-282a-4644-9aae-68f82c3ddf58-kube-api-access-xltwv\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.480037 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-config-data\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.480074 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d642739c-282a-4644-9aae-68f82c3ddf58-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.480105 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.480148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d642739c-282a-4644-9aae-68f82c3ddf58-logs\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.480204 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.482817 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.484057 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d642739c-282a-4644-9aae-68f82c3ddf58-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.485537 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d642739c-282a-4644-9aae-68f82c3ddf58-logs\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.486756 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-config-data\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.493344 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.493346 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.494314 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d642739c-282a-4644-9aae-68f82c3ddf58-scripts\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.506317 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xltwv\" (UniqueName: \"kubernetes.io/projected/d642739c-282a-4644-9aae-68f82c3ddf58-kube-api-access-xltwv\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.547028 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d642739c-282a-4644-9aae-68f82c3ddf58\") " pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.688964 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.825076 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.886646 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgq6n\" (UniqueName: \"kubernetes.io/projected/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-kube-api-access-kgq6n\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887061 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887091 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-config-data\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887117 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887194 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-logs\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887330 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-internal-tls-certs\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887351 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-scripts\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.887396 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-httpd-run\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.888581 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.888842 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-logs" (OuterVolumeSpecName: "logs") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.893031 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-kube-api-access-kgq6n" (OuterVolumeSpecName: "kube-api-access-kgq6n") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "kube-api-access-kgq6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.908688 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.908979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-scripts" (OuterVolumeSpecName: "scripts") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.989787 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.990605 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle\") pod \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\" (UID: \"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf\") " Nov 25 10:56:48 crc kubenswrapper[4696]: W1125 10:56:48.990790 4696 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf/volumes/kubernetes.io~secret/combined-ca-bundle Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.990877 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.992051 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.992135 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.992197 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgq6n\" (UniqueName: \"kubernetes.io/projected/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-kube-api-access-kgq6n\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.992594 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.992700 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:48 crc kubenswrapper[4696]: I1125 10:56:48.992784 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.016967 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.055373 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.094270 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-config-data" (OuterVolumeSpecName: "config-data") pod "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" (UID: "14c5ee2e-23a6-411e-9b8f-2cc34bc22daf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.094814 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.094827 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.094838 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.126833 4696 generic.go:334] "Generic (PLEG): container finished" podID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerID="84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae" exitCode=0 Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.126898 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf","Type":"ContainerDied","Data":"84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae"} Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.126923 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14c5ee2e-23a6-411e-9b8f-2cc34bc22daf","Type":"ContainerDied","Data":"b8ddae61627bc53cab526ce188526e19de95b865681eabea42bf9e6e3ce6268f"} Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.126943 4696 scope.go:117] "RemoveContainer" containerID="84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.127040 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.135077 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerStarted","Data":"513eb8ff774d08c16ce6d9eed0412781c30262b99f469caa9a77f43ba8de9ec4"} Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.173159 4696 scope.go:117] "RemoveContainer" containerID="541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.207060 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.221108 4696 scope.go:117] "RemoveContainer" containerID="84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae" Nov 25 10:56:49 crc kubenswrapper[4696]: E1125 10:56:49.221948 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae\": container with ID starting with 84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae not found: ID does not exist" containerID="84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.221973 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae"} err="failed to get container status \"84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae\": rpc error: code = NotFound desc = could not find container \"84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae\": container with ID starting with 84cda7d7367ba943dc1cfd029c6d0aff0d2e496057e8a75146c5e9d56ce522ae not found: ID does not exist" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.221993 4696 scope.go:117] "RemoveContainer" containerID="541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b" Nov 25 10:56:49 crc kubenswrapper[4696]: E1125 10:56:49.222207 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b\": container with ID starting with 541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b not found: ID does not exist" containerID="541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.222224 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b"} err="failed to get container status \"541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b\": rpc error: code = NotFound desc = could not find container \"541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b\": container with ID starting with 541f9c2676a1288b7b65d9ed6f903ff99af87c3db341a5515d18560b216ad80b not found: ID does not exist" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.258551 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.285442 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:56:49 crc kubenswrapper[4696]: E1125 10:56:49.285890 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-httpd" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.285903 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-httpd" Nov 25 10:56:49 crc kubenswrapper[4696]: E1125 10:56:49.285923 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-log" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.285929 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-log" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.286105 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-httpd" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.286123 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" containerName="glance-log" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.287011 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.296611 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.296802 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.312800 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402237 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402300 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrvfg\" (UniqueName: \"kubernetes.io/projected/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-kube-api-access-xrvfg\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402338 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402376 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402402 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402525 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.402552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.511031 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.511775 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.511830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.511915 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrvfg\" (UniqueName: \"kubernetes.io/projected/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-kube-api-access-xrvfg\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.511995 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.512083 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.512149 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.512312 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.512461 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.513380 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.513618 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.519893 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.526596 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.535530 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.537073 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.539296 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrvfg\" (UniqueName: \"kubernetes.io/projected/bd3e6e15-4a89-468d-8151-26c9e4c19dc5-kube-api-access-xrvfg\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.589256 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.614965 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd3e6e15-4a89-468d-8151-26c9e4c19dc5\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:56:49 crc kubenswrapper[4696]: I1125 10:56:49.621274 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.068314 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c5ee2e-23a6-411e-9b8f-2cc34bc22daf" path="/var/lib/kubelet/pods/14c5ee2e-23a6-411e-9b8f-2cc34bc22daf/volumes" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.070532 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42" path="/var/lib/kubelet/pods/ac5f4fea-e9ee-4fa9-b919-ba3a1bf21b42/volumes" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.185906 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d642739c-282a-4644-9aae-68f82c3ddf58","Type":"ContainerStarted","Data":"f3734b33e7274d970084d282c8194e3bbfa8ab6e0710d1c1451df58313844025"} Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.271286 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7w8x2"] Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.277602 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.322101 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.332347 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l5mz4" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.332650 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.372124 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7w8x2"] Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.437895 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.438015 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-config-data\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.438055 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4ljs\" (UniqueName: \"kubernetes.io/projected/35ce5f08-60ba-44bb-865c-b01c57ae49ac-kube-api-access-m4ljs\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.438091 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-scripts\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.539678 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-config-data\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.539930 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4ljs\" (UniqueName: \"kubernetes.io/projected/35ce5f08-60ba-44bb-865c-b01c57ae49ac-kube-api-access-m4ljs\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.540088 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-scripts\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.540227 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.546285 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-scripts\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.554963 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.562365 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-config-data\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.580983 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4ljs\" (UniqueName: \"kubernetes.io/projected/35ce5f08-60ba-44bb-865c-b01c57ae49ac-kube-api-access-m4ljs\") pod \"nova-cell0-conductor-db-sync-7w8x2\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.658397 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:56:50 crc kubenswrapper[4696]: I1125 10:56:50.950581 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:56:50 crc kubenswrapper[4696]: W1125 10:56:50.979463 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd3e6e15_4a89_468d_8151_26c9e4c19dc5.slice/crio-983087847e2a087bb4631df74ca8db630e465cbbcd85d63c38c00a6774d77cbd WatchSource:0}: Error finding container 983087847e2a087bb4631df74ca8db630e465cbbcd85d63c38c00a6774d77cbd: Status 404 returned error can't find the container with id 983087847e2a087bb4631df74ca8db630e465cbbcd85d63c38c00a6774d77cbd Nov 25 10:56:51 crc kubenswrapper[4696]: I1125 10:56:51.256310 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerStarted","Data":"86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb"} Nov 25 10:56:51 crc kubenswrapper[4696]: I1125 10:56:51.273712 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd3e6e15-4a89-468d-8151-26c9e4c19dc5","Type":"ContainerStarted","Data":"983087847e2a087bb4631df74ca8db630e465cbbcd85d63c38c00a6774d77cbd"} Nov 25 10:56:51 crc kubenswrapper[4696]: I1125 10:56:51.310493 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7w8x2"] Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.073828 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="5a8e77cd-7f3d-45e6-bb36-2ce208f70adb" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.170:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.301798 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd3e6e15-4a89-468d-8151-26c9e4c19dc5","Type":"ContainerStarted","Data":"cb9f47132865fca36fbf2a5b238cf1e5228b050c45f65ff9c87778367cdfe495"} Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.321787 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerStarted","Data":"dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0"} Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.323496 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" event={"ID":"35ce5f08-60ba-44bb-865c-b01c57ae49ac","Type":"ContainerStarted","Data":"7c29b459ab63523798f4c482ce4f7eafb970cefb8d559160a0fc6aec803fec62"} Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.325964 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d642739c-282a-4644-9aae-68f82c3ddf58","Type":"ContainerStarted","Data":"e6363395aca5e049caa2a42ae4300bb9117c57866936b51e830e11bcbafa6502"} Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.771148 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:56:52 crc kubenswrapper[4696]: I1125 10:56:52.879327 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6558fc797b-2qp4r" podUID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 10:56:53 crc kubenswrapper[4696]: I1125 10:56:53.094961 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="5a8e77cd-7f3d-45e6-bb36-2ce208f70adb" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.170:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:56:53 crc kubenswrapper[4696]: I1125 10:56:53.340692 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d642739c-282a-4644-9aae-68f82c3ddf58","Type":"ContainerStarted","Data":"f38cc09831cfc840f592e171d4b8b3620486cc8226df4a9139a521c9098dc6ae"} Nov 25 10:56:53 crc kubenswrapper[4696]: I1125 10:56:53.344843 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd3e6e15-4a89-468d-8151-26c9e4c19dc5","Type":"ContainerStarted","Data":"aa060904fdc2ade352cd98e9be51f0021cf3aa27630a502117e1855704443e90"} Nov 25 10:56:53 crc kubenswrapper[4696]: I1125 10:56:53.351167 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerStarted","Data":"3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5"} Nov 25 10:56:53 crc kubenswrapper[4696]: I1125 10:56:53.364112 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.364093521 podStartE2EDuration="5.364093521s" podCreationTimestamp="2025-11-25 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:53.360882118 +0000 UTC m=+1210.174498695" watchObservedRunningTime="2025-11-25 10:56:53.364093521 +0000 UTC m=+1210.177710108" Nov 25 10:56:53 crc kubenswrapper[4696]: I1125 10:56:53.406438 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.406417253 podStartE2EDuration="4.406417253s" podCreationTimestamp="2025-11-25 10:56:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:56:53.384852863 +0000 UTC m=+1210.198469450" watchObservedRunningTime="2025-11-25 10:56:53.406417253 +0000 UTC m=+1210.220033840" Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.389727 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerStarted","Data":"19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3"} Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.390013 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-central-agent" containerID="cri-o://86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb" gracePeriod=30 Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.390178 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.390286 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="proxy-httpd" containerID="cri-o://19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3" gracePeriod=30 Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.390392 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="sg-core" containerID="cri-o://3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5" gracePeriod=30 Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.390446 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-notification-agent" containerID="cri-o://dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0" gracePeriod=30 Nov 25 10:56:55 crc kubenswrapper[4696]: I1125 10:56:55.423395 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.467479589 podStartE2EDuration="8.423378356s" podCreationTimestamp="2025-11-25 10:56:47 +0000 UTC" firstStartedPulling="2025-11-25 10:56:48.223602908 +0000 UTC m=+1205.037219485" lastFinishedPulling="2025-11-25 10:56:54.179501665 +0000 UTC m=+1210.993118252" observedRunningTime="2025-11-25 10:56:55.419253285 +0000 UTC m=+1212.232869872" watchObservedRunningTime="2025-11-25 10:56:55.423378356 +0000 UTC m=+1212.236994943" Nov 25 10:56:56 crc kubenswrapper[4696]: I1125 10:56:56.403333 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerDied","Data":"19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3"} Nov 25 10:56:56 crc kubenswrapper[4696]: I1125 10:56:56.403288 4696 generic.go:334] "Generic (PLEG): container finished" podID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerID="19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3" exitCode=0 Nov 25 10:56:56 crc kubenswrapper[4696]: I1125 10:56:56.403727 4696 generic.go:334] "Generic (PLEG): container finished" podID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerID="3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5" exitCode=2 Nov 25 10:56:56 crc kubenswrapper[4696]: I1125 10:56:56.403739 4696 generic.go:334] "Generic (PLEG): container finished" podID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerID="dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0" exitCode=0 Nov 25 10:56:56 crc kubenswrapper[4696]: I1125 10:56:56.403755 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerDied","Data":"3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5"} Nov 25 10:56:56 crc kubenswrapper[4696]: I1125 10:56:56.403765 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerDied","Data":"dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0"} Nov 25 10:56:57 crc kubenswrapper[4696]: I1125 10:56:57.731392 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 10:56:58 crc kubenswrapper[4696]: I1125 10:56:58.825508 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:56:58 crc kubenswrapper[4696]: I1125 10:56:58.825768 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:56:58 crc kubenswrapper[4696]: I1125 10:56:58.872957 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:56:58 crc kubenswrapper[4696]: I1125 10:56:58.896069 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:56:59 crc kubenswrapper[4696]: I1125 10:56:59.455233 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:56:59 crc kubenswrapper[4696]: I1125 10:56:59.455525 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:56:59 crc kubenswrapper[4696]: I1125 10:56:59.622293 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:59 crc kubenswrapper[4696]: I1125 10:56:59.622442 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:59 crc kubenswrapper[4696]: I1125 10:56:59.699483 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:56:59 crc kubenswrapper[4696]: I1125 10:56:59.702900 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:57:00 crc kubenswrapper[4696]: I1125 10:57:00.470007 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:57:00 crc kubenswrapper[4696]: I1125 10:57:00.470049 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:57:00 crc kubenswrapper[4696]: I1125 10:57:00.801521 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:57:00 crc kubenswrapper[4696]: I1125 10:57:00.801569 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.260739 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.261084 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.267676 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.489208 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.489394 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.762909 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:57:02 crc kubenswrapper[4696]: I1125 10:57:02.879125 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6558fc797b-2qp4r" podUID="1944527b-6ffc-4b6b-954f-4b01394ea0cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 10:57:04 crc kubenswrapper[4696]: I1125 10:57:04.381221 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:57:04 crc kubenswrapper[4696]: I1125 10:57:04.381586 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:57:04 crc kubenswrapper[4696]: I1125 10:57:04.384441 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:57:05 crc kubenswrapper[4696]: I1125 10:57:05.529718 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" event={"ID":"35ce5f08-60ba-44bb-865c-b01c57ae49ac","Type":"ContainerStarted","Data":"3c7adebf28f9d5cf798a7e088bc7cc4ca91ecfcdc49dbb27cb1727dcee908ef5"} Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.165615 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.195191 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" podStartSLOduration=6.696154681 podStartE2EDuration="20.195169004s" podCreationTimestamp="2025-11-25 10:56:50 +0000 UTC" firstStartedPulling="2025-11-25 10:56:51.377825297 +0000 UTC m=+1208.191441884" lastFinishedPulling="2025-11-25 10:57:04.87683962 +0000 UTC m=+1221.690456207" observedRunningTime="2025-11-25 10:57:05.553091861 +0000 UTC m=+1222.366708468" watchObservedRunningTime="2025-11-25 10:57:10.195169004 +0000 UTC m=+1227.008785591" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279276 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-config-data\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279373 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdt5p\" (UniqueName: \"kubernetes.io/projected/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-kube-api-access-fdt5p\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279407 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-combined-ca-bundle\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279429 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-log-httpd\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279464 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-run-httpd\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279486 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-scripts\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.279596 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-sg-core-conf-yaml\") pod \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\" (UID: \"3cb97af4-ede7-4b3a-beb1-30fe974a3dca\") " Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.281455 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.281498 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.299254 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-kube-api-access-fdt5p" (OuterVolumeSpecName: "kube-api-access-fdt5p") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "kube-api-access-fdt5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.314699 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-scripts" (OuterVolumeSpecName: "scripts") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.334168 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.382153 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdt5p\" (UniqueName: \"kubernetes.io/projected/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-kube-api-access-fdt5p\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.382386 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.382453 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.382528 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.382586 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.403299 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.410861 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-config-data" (OuterVolumeSpecName: "config-data") pod "3cb97af4-ede7-4b3a-beb1-30fe974a3dca" (UID: "3cb97af4-ede7-4b3a-beb1-30fe974a3dca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.483673 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.483701 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb97af4-ede7-4b3a-beb1-30fe974a3dca-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.578956 4696 generic.go:334] "Generic (PLEG): container finished" podID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerID="86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb" exitCode=0 Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.579002 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.579020 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerDied","Data":"86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb"} Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.579528 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cb97af4-ede7-4b3a-beb1-30fe974a3dca","Type":"ContainerDied","Data":"513eb8ff774d08c16ce6d9eed0412781c30262b99f469caa9a77f43ba8de9ec4"} Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.579601 4696 scope.go:117] "RemoveContainer" containerID="19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.614907 4696 scope.go:117] "RemoveContainer" containerID="3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.638366 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.645062 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.656227 4696 scope.go:117] "RemoveContainer" containerID="dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.677124 4696 scope.go:117] "RemoveContainer" containerID="86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.684332 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.684733 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="proxy-httpd" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.684748 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="proxy-httpd" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.684757 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="sg-core" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.684765 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="sg-core" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.684780 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-central-agent" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.684786 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-central-agent" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.684800 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-notification-agent" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.684806 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-notification-agent" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.684977 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="sg-core" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.685004 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="proxy-httpd" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.685012 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-notification-agent" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.685022 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" containerName="ceilometer-central-agent" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.686527 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.692085 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.699350 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.699833 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.704973 4696 scope.go:117] "RemoveContainer" containerID="19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.708722 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3\": container with ID starting with 19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3 not found: ID does not exist" containerID="19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.708761 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3"} err="failed to get container status \"19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3\": rpc error: code = NotFound desc = could not find container \"19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3\": container with ID starting with 19ebc141e6d2b063813a55224588c363edfcb833a690e0eca894f7b257ee58c3 not found: ID does not exist" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.708787 4696 scope.go:117] "RemoveContainer" containerID="3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.710249 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5\": container with ID starting with 3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5 not found: ID does not exist" containerID="3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.710293 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5"} err="failed to get container status \"3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5\": rpc error: code = NotFound desc = could not find container \"3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5\": container with ID starting with 3dfc5b26d212d62bf73af743d99b4985057170e8a14fb38b97d3586946b5a1d5 not found: ID does not exist" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.710310 4696 scope.go:117] "RemoveContainer" containerID="dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.711182 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0\": container with ID starting with dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0 not found: ID does not exist" containerID="dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.711199 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0"} err="failed to get container status \"dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0\": rpc error: code = NotFound desc = could not find container \"dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0\": container with ID starting with dcb3214aeb4a8b80c7c8d430691a4b9e231b06ace6b4e622295515bf9f3ea7a0 not found: ID does not exist" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.711229 4696 scope.go:117] "RemoveContainer" containerID="86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb" Nov 25 10:57:10 crc kubenswrapper[4696]: E1125 10:57:10.714933 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb\": container with ID starting with 86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb not found: ID does not exist" containerID="86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.714975 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb"} err="failed to get container status \"86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb\": rpc error: code = NotFound desc = could not find container \"86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb\": container with ID starting with 86504c07c92b25bd4e7ba1235ad58254230acb77a71517fbc68706a16fe446cb not found: ID does not exist" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.787501 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-scripts\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.787786 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.787915 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-log-httpd\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.788009 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-run-httpd\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.788140 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-config-data\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.788233 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.788323 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h8x7\" (UniqueName: \"kubernetes.io/projected/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-kube-api-access-7h8x7\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890157 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890215 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h8x7\" (UniqueName: \"kubernetes.io/projected/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-kube-api-access-7h8x7\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890259 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-scripts\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890290 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890341 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-log-httpd\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890372 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-run-httpd\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.890436 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-config-data\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.891419 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-log-httpd\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.891424 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-run-httpd\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.895393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.896023 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-config-data\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.896423 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-scripts\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.898179 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:10 crc kubenswrapper[4696]: I1125 10:57:10.908811 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h8x7\" (UniqueName: \"kubernetes.io/projected/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-kube-api-access-7h8x7\") pod \"ceilometer-0\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " pod="openstack/ceilometer-0" Nov 25 10:57:11 crc kubenswrapper[4696]: I1125 10:57:11.019339 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:11 crc kubenswrapper[4696]: I1125 10:57:11.516976 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:11 crc kubenswrapper[4696]: I1125 10:57:11.618980 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerStarted","Data":"81de1837540ebb8ce8b2c8e9d68bd74e558276860c61e7146e99ef9862bb06aa"} Nov 25 10:57:12 crc kubenswrapper[4696]: I1125 10:57:12.055151 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb97af4-ede7-4b3a-beb1-30fe974a3dca" path="/var/lib/kubelet/pods/3cb97af4-ede7-4b3a-beb1-30fe974a3dca/volumes" Nov 25 10:57:12 crc kubenswrapper[4696]: I1125 10:57:12.629732 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerStarted","Data":"47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5"} Nov 25 10:57:13 crc kubenswrapper[4696]: I1125 10:57:13.639008 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerStarted","Data":"e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f"} Nov 25 10:57:14 crc kubenswrapper[4696]: I1125 10:57:14.650961 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerStarted","Data":"20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71"} Nov 25 10:57:15 crc kubenswrapper[4696]: I1125 10:57:15.660564 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerStarted","Data":"63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487"} Nov 25 10:57:15 crc kubenswrapper[4696]: I1125 10:57:15.661090 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:57:15 crc kubenswrapper[4696]: I1125 10:57:15.679993 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.509844926 podStartE2EDuration="5.679978214s" podCreationTimestamp="2025-11-25 10:57:10 +0000 UTC" firstStartedPulling="2025-11-25 10:57:11.522555901 +0000 UTC m=+1228.336172488" lastFinishedPulling="2025-11-25 10:57:14.692689189 +0000 UTC m=+1231.506305776" observedRunningTime="2025-11-25 10:57:15.678416364 +0000 UTC m=+1232.492032951" watchObservedRunningTime="2025-11-25 10:57:15.679978214 +0000 UTC m=+1232.493594801" Nov 25 10:57:16 crc kubenswrapper[4696]: I1125 10:57:16.224114 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:57:16 crc kubenswrapper[4696]: I1125 10:57:16.293330 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.223458 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.314320 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6558fc797b-2qp4r" Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.392619 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c8fd89c8d-5vb5f"] Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.687272 4696 generic.go:334] "Generic (PLEG): container finished" podID="35ce5f08-60ba-44bb-865c-b01c57ae49ac" containerID="3c7adebf28f9d5cf798a7e088bc7cc4ca91ecfcdc49dbb27cb1727dcee908ef5" exitCode=0 Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.687517 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon-log" containerID="cri-o://2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154" gracePeriod=30 Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.687883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" event={"ID":"35ce5f08-60ba-44bb-865c-b01c57ae49ac","Type":"ContainerDied","Data":"3c7adebf28f9d5cf798a7e088bc7cc4ca91ecfcdc49dbb27cb1727dcee908ef5"} Nov 25 10:57:18 crc kubenswrapper[4696]: I1125 10:57:18.688376 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" containerID="cri-o://91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829" gracePeriod=30 Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.010147 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.055592 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4ljs\" (UniqueName: \"kubernetes.io/projected/35ce5f08-60ba-44bb-865c-b01c57ae49ac-kube-api-access-m4ljs\") pod \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.055762 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-config-data\") pod \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.055818 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-scripts\") pod \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.055868 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-combined-ca-bundle\") pod \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\" (UID: \"35ce5f08-60ba-44bb-865c-b01c57ae49ac\") " Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.061218 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ce5f08-60ba-44bb-865c-b01c57ae49ac-kube-api-access-m4ljs" (OuterVolumeSpecName: "kube-api-access-m4ljs") pod "35ce5f08-60ba-44bb-865c-b01c57ae49ac" (UID: "35ce5f08-60ba-44bb-865c-b01c57ae49ac"). InnerVolumeSpecName "kube-api-access-m4ljs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.083305 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-scripts" (OuterVolumeSpecName: "scripts") pod "35ce5f08-60ba-44bb-865c-b01c57ae49ac" (UID: "35ce5f08-60ba-44bb-865c-b01c57ae49ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.091513 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-config-data" (OuterVolumeSpecName: "config-data") pod "35ce5f08-60ba-44bb-865c-b01c57ae49ac" (UID: "35ce5f08-60ba-44bb-865c-b01c57ae49ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.091631 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35ce5f08-60ba-44bb-865c-b01c57ae49ac" (UID: "35ce5f08-60ba-44bb-865c-b01c57ae49ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.158889 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.158920 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.158929 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ce5f08-60ba-44bb-865c-b01c57ae49ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.158940 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4ljs\" (UniqueName: \"kubernetes.io/projected/35ce5f08-60ba-44bb-865c-b01c57ae49ac-kube-api-access-m4ljs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.705724 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" event={"ID":"35ce5f08-60ba-44bb-865c-b01c57ae49ac","Type":"ContainerDied","Data":"7c29b459ab63523798f4c482ce4f7eafb970cefb8d559160a0fc6aec803fec62"} Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.705767 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c29b459ab63523798f4c482ce4f7eafb970cefb8d559160a0fc6aec803fec62" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.705830 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7w8x2" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.814421 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:57:20 crc kubenswrapper[4696]: E1125 10:57:20.814874 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ce5f08-60ba-44bb-865c-b01c57ae49ac" containerName="nova-cell0-conductor-db-sync" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.814905 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ce5f08-60ba-44bb-865c-b01c57ae49ac" containerName="nova-cell0-conductor-db-sync" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.815123 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ce5f08-60ba-44bb-865c-b01c57ae49ac" containerName="nova-cell0-conductor-db-sync" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.815703 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.817758 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l5mz4" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.819336 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.832594 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.870905 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8189ed6c-e99d-44db-bd26-7b74ea6539fb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.871015 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8189ed6c-e99d-44db-bd26-7b74ea6539fb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.871047 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7jw7\" (UniqueName: \"kubernetes.io/projected/8189ed6c-e99d-44db-bd26-7b74ea6539fb-kube-api-access-h7jw7\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.973084 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8189ed6c-e99d-44db-bd26-7b74ea6539fb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.973186 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8189ed6c-e99d-44db-bd26-7b74ea6539fb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.973234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7jw7\" (UniqueName: \"kubernetes.io/projected/8189ed6c-e99d-44db-bd26-7b74ea6539fb-kube-api-access-h7jw7\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.976777 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8189ed6c-e99d-44db-bd26-7b74ea6539fb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.984726 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8189ed6c-e99d-44db-bd26-7b74ea6539fb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:20 crc kubenswrapper[4696]: I1125 10:57:20.998960 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7jw7\" (UniqueName: \"kubernetes.io/projected/8189ed6c-e99d-44db-bd26-7b74ea6539fb-kube-api-access-h7jw7\") pod \"nova-cell0-conductor-0\" (UID: \"8189ed6c-e99d-44db-bd26-7b74ea6539fb\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:21 crc kubenswrapper[4696]: I1125 10:57:21.132950 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:21 crc kubenswrapper[4696]: I1125 10:57:21.559469 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:57:21 crc kubenswrapper[4696]: I1125 10:57:21.714707 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8189ed6c-e99d-44db-bd26-7b74ea6539fb","Type":"ContainerStarted","Data":"c9e1bdb51987ecee29081ad4984939318aab52b9f3f799dec329f5ad7e81bb38"} Nov 25 10:57:22 crc kubenswrapper[4696]: I1125 10:57:22.723127 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8189ed6c-e99d-44db-bd26-7b74ea6539fb","Type":"ContainerStarted","Data":"efdf47d72ea9c56969b62718c4b59693b3637f8f449ce81f86b03ebacaf919f2"} Nov 25 10:57:22 crc kubenswrapper[4696]: I1125 10:57:22.723608 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:22 crc kubenswrapper[4696]: I1125 10:57:22.726429 4696 generic.go:334] "Generic (PLEG): container finished" podID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerID="91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829" exitCode=0 Nov 25 10:57:22 crc kubenswrapper[4696]: I1125 10:57:22.726473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerDied","Data":"91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829"} Nov 25 10:57:22 crc kubenswrapper[4696]: I1125 10:57:22.726504 4696 scope.go:117] "RemoveContainer" containerID="b743c34bdb2c734c6424f22dad595eb14070e34899fbea0e6b459832b98e3148" Nov 25 10:57:22 crc kubenswrapper[4696]: I1125 10:57:22.761922 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.166181 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.187625 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=6.187547095 podStartE2EDuration="6.187547095s" podCreationTimestamp="2025-11-25 10:57:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:22.739179096 +0000 UTC m=+1239.552795683" watchObservedRunningTime="2025-11-25 10:57:26.187547095 +0000 UTC m=+1243.001163702" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.605113 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lfvg7"] Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.606911 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.611487 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.611587 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.621045 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfvg7"] Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.671784 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.672027 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqf8m\" (UniqueName: \"kubernetes.io/projected/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-kube-api-access-dqf8m\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.672120 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-scripts\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.672344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-config-data\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.775607 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqf8m\" (UniqueName: \"kubernetes.io/projected/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-kube-api-access-dqf8m\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.775695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-scripts\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.775803 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-config-data\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.775839 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.857159 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.860088 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-config-data\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.860253 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.861861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqf8m\" (UniqueName: \"kubernetes.io/projected/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-kube-api-access-dqf8m\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.872218 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-scripts\") pod \"nova-cell0-cell-mapping-lfvg7\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.874279 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.880179 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.907861 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-config-data\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.908126 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.908409 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlqv2\" (UniqueName: \"kubernetes.io/projected/e7a7d178-8896-46b6-a2c6-b5530d20402e-kube-api-access-rlqv2\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.908536 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a7d178-8896-46b6-a2c6-b5530d20402e-logs\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.937687 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:26 crc kubenswrapper[4696]: I1125 10:57:26.949062 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.000374 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.003228 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.021027 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.023443 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-config-data\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.023695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.023938 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlqv2\" (UniqueName: \"kubernetes.io/projected/e7a7d178-8896-46b6-a2c6-b5530d20402e-kube-api-access-rlqv2\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.024087 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a7d178-8896-46b6-a2c6-b5530d20402e-logs\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.024226 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-config-data\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.024359 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pntt4\" (UniqueName: \"kubernetes.io/projected/8a59b88b-c4ab-42cb-81df-8baaabbae785-kube-api-access-pntt4\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.026185 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.033620 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a7d178-8896-46b6-a2c6-b5530d20402e-logs\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.052488 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-config-data\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.065250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.074556 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.086597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlqv2\" (UniqueName: \"kubernetes.io/projected/e7a7d178-8896-46b6-a2c6-b5530d20402e-kube-api-access-rlqv2\") pod \"nova-api-0\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.129228 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.129830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-config-data\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.129951 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pntt4\" (UniqueName: \"kubernetes.io/projected/8a59b88b-c4ab-42cb-81df-8baaabbae785-kube-api-access-pntt4\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.153058 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-config-data\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.157198 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.167821 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.170186 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.179109 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.200884 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pntt4\" (UniqueName: \"kubernetes.io/projected/8a59b88b-c4ab-42cb-81df-8baaabbae785-kube-api-access-pntt4\") pod \"nova-scheduler-0\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.236596 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-config-data\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.236674 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-logs\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.236716 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.236782 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52ww6\" (UniqueName: \"kubernetes.io/projected/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-kube-api-access-52ww6\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.253807 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.260313 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.344040 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-config-data\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.344106 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-logs\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.344156 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.344234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52ww6\" (UniqueName: \"kubernetes.io/projected/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-kube-api-access-52ww6\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.347526 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-logs\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.361648 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-config-data\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.363475 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.382528 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.387148 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.392176 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.392890 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.408849 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.415119 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52ww6\" (UniqueName: \"kubernetes.io/projected/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-kube-api-access-52ww6\") pod \"nova-metadata-0\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.446632 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4wvx4"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.455194 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.455263 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.455285 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56z99\" (UniqueName: \"kubernetes.io/projected/23fb3f29-8c86-4ba7-8260-4ead78279db6-kube-api-access-56z99\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.465491 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.503585 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.528684 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4wvx4"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.557617 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.557703 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.557735 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56z99\" (UniqueName: \"kubernetes.io/projected/23fb3f29-8c86-4ba7-8260-4ead78279db6-kube-api-access-56z99\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.571987 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.575153 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.583273 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56z99\" (UniqueName: \"kubernetes.io/projected/23fb3f29-8c86-4ba7-8260-4ead78279db6-kube-api-access-56z99\") pod \"nova-cell1-novncproxy-0\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.661866 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-config\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.661898 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfwft\" (UniqueName: \"kubernetes.io/projected/15f521d2-8aef-4ea6-b26d-1e2a259477ee-kube-api-access-qfwft\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.661955 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-svc\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.661991 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.662106 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.662130 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.718163 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.763305 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.763416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.763451 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.763497 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-config\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.763517 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfwft\" (UniqueName: \"kubernetes.io/projected/15f521d2-8aef-4ea6-b26d-1e2a259477ee-kube-api-access-qfwft\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.763556 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-svc\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.764441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-svc\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.764995 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.765108 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-config\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.765522 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.765766 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfvg7"] Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.766028 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.796621 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfwft\" (UniqueName: \"kubernetes.io/projected/15f521d2-8aef-4ea6-b26d-1e2a259477ee-kube-api-access-qfwft\") pod \"dnsmasq-dns-bccf8f775-4wvx4\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.812490 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:27 crc kubenswrapper[4696]: I1125 10:57:27.865073 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfvg7" event={"ID":"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527","Type":"ContainerStarted","Data":"7be4dfae74b41fd675f56d71ff5fe142941015d7a6535cddc5db3f6ea6550c24"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.068971 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:28 crc kubenswrapper[4696]: W1125 10:57:28.078054 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7a7d178_8896_46b6_a2c6_b5530d20402e.slice/crio-950f19bb5acfceef768a030d4223ba6081b4c59a2fe39d6981d068642328586d WatchSource:0}: Error finding container 950f19bb5acfceef768a030d4223ba6081b4c59a2fe39d6981d068642328586d: Status 404 returned error can't find the container with id 950f19bb5acfceef768a030d4223ba6081b4c59a2fe39d6981d068642328586d Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.302284 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:28 crc kubenswrapper[4696]: W1125 10:57:28.304102 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a59b88b_c4ab_42cb_81df_8baaabbae785.slice/crio-9bea28ce411633618cf456894816714d588200a056751677f307a25d48e09abf WatchSource:0}: Error finding container 9bea28ce411633618cf456894816714d588200a056751677f307a25d48e09abf: Status 404 returned error can't find the container with id 9bea28ce411633618cf456894816714d588200a056751677f307a25d48e09abf Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.443737 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:28 crc kubenswrapper[4696]: W1125 10:57:28.454596 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38c0c4b1_97f2_4bec_bba5_df9fe04663ac.slice/crio-a6ec50d7c19f3532e95f6da9fc6f64a7fb4f88a2f2115befaacf877ebcd5a01f WatchSource:0}: Error finding container a6ec50d7c19f3532e95f6da9fc6f64a7fb4f88a2f2115befaacf877ebcd5a01f: Status 404 returned error can't find the container with id a6ec50d7c19f3532e95f6da9fc6f64a7fb4f88a2f2115befaacf877ebcd5a01f Nov 25 10:57:28 crc kubenswrapper[4696]: W1125 10:57:28.549545 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23fb3f29_8c86_4ba7_8260_4ead78279db6.slice/crio-34674ceea1ab5f008225bda043843b18c0d4253a513a34b884821e553e04e7cc WatchSource:0}: Error finding container 34674ceea1ab5f008225bda043843b18c0d4253a513a34b884821e553e04e7cc: Status 404 returned error can't find the container with id 34674ceea1ab5f008225bda043843b18c0d4253a513a34b884821e553e04e7cc Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.550972 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.560709 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4wvx4"] Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.832376 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfdng"] Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.833939 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.839542 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfdng"] Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.841214 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.841332 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.897586 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7a7d178-8896-46b6-a2c6-b5530d20402e","Type":"ContainerStarted","Data":"950f19bb5acfceef768a030d4223ba6081b4c59a2fe39d6981d068642328586d"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.899726 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a59b88b-c4ab-42cb-81df-8baaabbae785","Type":"ContainerStarted","Data":"9bea28ce411633618cf456894816714d588200a056751677f307a25d48e09abf"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.901513 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38c0c4b1-97f2-4bec-bba5-df9fe04663ac","Type":"ContainerStarted","Data":"a6ec50d7c19f3532e95f6da9fc6f64a7fb4f88a2f2115befaacf877ebcd5a01f"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.907338 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" event={"ID":"15f521d2-8aef-4ea6-b26d-1e2a259477ee","Type":"ContainerStarted","Data":"90ec1f9ca882c89ea4e00222a652d851f458daa620b7d64bc02c9c41d62b5406"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.910495 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"23fb3f29-8c86-4ba7-8260-4ead78279db6","Type":"ContainerStarted","Data":"34674ceea1ab5f008225bda043843b18c0d4253a513a34b884821e553e04e7cc"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.913700 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfvg7" event={"ID":"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527","Type":"ContainerStarted","Data":"96e7cd861737e0e9581db9df692210348ae8c3c1c31ea2adea0c671058b1f2c5"} Nov 25 10:57:28 crc kubenswrapper[4696]: I1125 10:57:28.973880 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lfvg7" podStartSLOduration=2.973852707 podStartE2EDuration="2.973852707s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:28.924814845 +0000 UTC m=+1245.738431432" watchObservedRunningTime="2025-11-25 10:57:28.973852707 +0000 UTC m=+1245.787469294" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.003883 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.009924 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-config-data\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.009971 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bdkq\" (UniqueName: \"kubernetes.io/projected/4b10ca5a-c03c-407e-9d19-c649df1f9711-kube-api-access-2bdkq\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.010020 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-scripts\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.113118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.114100 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-config-data\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.114129 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bdkq\" (UniqueName: \"kubernetes.io/projected/4b10ca5a-c03c-407e-9d19-c649df1f9711-kube-api-access-2bdkq\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.114162 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-scripts\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.128244 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-config-data\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.138404 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.144268 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bdkq\" (UniqueName: \"kubernetes.io/projected/4b10ca5a-c03c-407e-9d19-c649df1f9711-kube-api-access-2bdkq\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.146747 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-scripts\") pod \"nova-cell1-conductor-db-sync-xfdng\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.180231 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.750580 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfdng"] Nov 25 10:57:29 crc kubenswrapper[4696]: W1125 10:57:29.764027 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b10ca5a_c03c_407e_9d19_c649df1f9711.slice/crio-a6da4be66c9513f840d851b033368a92d9f41e7786c30b217578f64f8a9b5257 WatchSource:0}: Error finding container a6da4be66c9513f840d851b033368a92d9f41e7786c30b217578f64f8a9b5257: Status 404 returned error can't find the container with id a6da4be66c9513f840d851b033368a92d9f41e7786c30b217578f64f8a9b5257 Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.937092 4696 generic.go:334] "Generic (PLEG): container finished" podID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerID="bceb37e2683d4e9ee659c7172a432555d7927e553779cdaced918cb5dbad1b2f" exitCode=0 Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.937232 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" event={"ID":"15f521d2-8aef-4ea6-b26d-1e2a259477ee","Type":"ContainerDied","Data":"bceb37e2683d4e9ee659c7172a432555d7927e553779cdaced918cb5dbad1b2f"} Nov 25 10:57:29 crc kubenswrapper[4696]: I1125 10:57:29.941567 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfdng" event={"ID":"4b10ca5a-c03c-407e-9d19-c649df1f9711","Type":"ContainerStarted","Data":"a6da4be66c9513f840d851b033368a92d9f41e7786c30b217578f64f8a9b5257"} Nov 25 10:57:30 crc kubenswrapper[4696]: I1125 10:57:30.802271 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:57:30 crc kubenswrapper[4696]: I1125 10:57:30.802524 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:57:31 crc kubenswrapper[4696]: I1125 10:57:31.364040 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:57:31 crc kubenswrapper[4696]: I1125 10:57:31.373861 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.761326 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.972396 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"23fb3f29-8c86-4ba7-8260-4ead78279db6","Type":"ContainerStarted","Data":"3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.972513 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="23fb3f29-8c86-4ba7-8260-4ead78279db6" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b" gracePeriod=30 Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.975005 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7a7d178-8896-46b6-a2c6-b5530d20402e","Type":"ContainerStarted","Data":"14b71c8f64593d8205994f62562a3b8fd60a91a3c97ed9cdd437ad5ad84083b0"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.975040 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7a7d178-8896-46b6-a2c6-b5530d20402e","Type":"ContainerStarted","Data":"73fe6adabb90bd75e23a61242f35e20e563055adcaf260fcb9e3176f5e24f6df"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.980383 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a59b88b-c4ab-42cb-81df-8baaabbae785","Type":"ContainerStarted","Data":"b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.982750 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38c0c4b1-97f2-4bec-bba5-df9fe04663ac","Type":"ContainerStarted","Data":"d9265fa05182b7a257b182f89adfc9802f8bf1f422914bbd848c5aca90cf2827"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.982790 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38c0c4b1-97f2-4bec-bba5-df9fe04663ac","Type":"ContainerStarted","Data":"ce495edb919adc1a70e8680ce556870be7891cd6340328035a4df6886ea57703"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.983005 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-log" containerID="cri-o://ce495edb919adc1a70e8680ce556870be7891cd6340328035a4df6886ea57703" gracePeriod=30 Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.983033 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-metadata" containerID="cri-o://d9265fa05182b7a257b182f89adfc9802f8bf1f422914bbd848c5aca90cf2827" gracePeriod=30 Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.985379 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfdng" event={"ID":"4b10ca5a-c03c-407e-9d19-c649df1f9711","Type":"ContainerStarted","Data":"4d316d2a2a1f34711da8f0a58a99152d05248979396e27f75374e53e5f17b288"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.993594 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.788011238 podStartE2EDuration="5.993576702s" podCreationTimestamp="2025-11-25 10:57:27 +0000 UTC" firstStartedPulling="2025-11-25 10:57:28.552489534 +0000 UTC m=+1245.366106121" lastFinishedPulling="2025-11-25 10:57:31.758054998 +0000 UTC m=+1248.571671585" observedRunningTime="2025-11-25 10:57:32.991114328 +0000 UTC m=+1249.804730935" watchObservedRunningTime="2025-11-25 10:57:32.993576702 +0000 UTC m=+1249.807193289" Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.995652 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" event={"ID":"15f521d2-8aef-4ea6-b26d-1e2a259477ee","Type":"ContainerStarted","Data":"30810c9e3eecbf9e1be83ee7ada1ac9b5ccadf967e125fb72e9068e4d53c64f7"} Nov 25 10:57:32 crc kubenswrapper[4696]: I1125 10:57:32.996033 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:33 crc kubenswrapper[4696]: I1125 10:57:33.014282 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.715739674 podStartE2EDuration="6.014264827s" podCreationTimestamp="2025-11-25 10:57:27 +0000 UTC" firstStartedPulling="2025-11-25 10:57:28.457293781 +0000 UTC m=+1245.270910368" lastFinishedPulling="2025-11-25 10:57:31.755818934 +0000 UTC m=+1248.569435521" observedRunningTime="2025-11-25 10:57:33.008709246 +0000 UTC m=+1249.822325833" watchObservedRunningTime="2025-11-25 10:57:33.014264827 +0000 UTC m=+1249.827881404" Nov 25 10:57:33 crc kubenswrapper[4696]: I1125 10:57:33.036241 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xfdng" podStartSLOduration=5.03621548 podStartE2EDuration="5.03621548s" podCreationTimestamp="2025-11-25 10:57:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:33.026260822 +0000 UTC m=+1249.839877409" watchObservedRunningTime="2025-11-25 10:57:33.03621548 +0000 UTC m=+1249.849832067" Nov 25 10:57:33 crc kubenswrapper[4696]: I1125 10:57:33.083984 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.409265603 podStartE2EDuration="7.083965433s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="2025-11-25 10:57:28.080377696 +0000 UTC m=+1244.893994283" lastFinishedPulling="2025-11-25 10:57:31.755077526 +0000 UTC m=+1248.568694113" observedRunningTime="2025-11-25 10:57:33.062522468 +0000 UTC m=+1249.876139055" watchObservedRunningTime="2025-11-25 10:57:33.083965433 +0000 UTC m=+1249.897582020" Nov 25 10:57:33 crc kubenswrapper[4696]: I1125 10:57:33.085450 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.636145333 podStartE2EDuration="7.085440498s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="2025-11-25 10:57:28.306512168 +0000 UTC m=+1245.120128755" lastFinishedPulling="2025-11-25 10:57:31.755807333 +0000 UTC m=+1248.569423920" observedRunningTime="2025-11-25 10:57:33.080144707 +0000 UTC m=+1249.893761294" watchObservedRunningTime="2025-11-25 10:57:33.085440498 +0000 UTC m=+1249.899057095" Nov 25 10:57:33 crc kubenswrapper[4696]: I1125 10:57:33.126172 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" podStartSLOduration=6.126152844 podStartE2EDuration="6.126152844s" podCreationTimestamp="2025-11-25 10:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:33.121954224 +0000 UTC m=+1249.935570821" watchObservedRunningTime="2025-11-25 10:57:33.126152844 +0000 UTC m=+1249.939769431" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.096670 4696 generic.go:334] "Generic (PLEG): container finished" podID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerID="d9265fa05182b7a257b182f89adfc9802f8bf1f422914bbd848c5aca90cf2827" exitCode=0 Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.096923 4696 generic.go:334] "Generic (PLEG): container finished" podID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerID="ce495edb919adc1a70e8680ce556870be7891cd6340328035a4df6886ea57703" exitCode=143 Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.097739 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38c0c4b1-97f2-4bec-bba5-df9fe04663ac","Type":"ContainerDied","Data":"d9265fa05182b7a257b182f89adfc9802f8bf1f422914bbd848c5aca90cf2827"} Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.097765 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38c0c4b1-97f2-4bec-bba5-df9fe04663ac","Type":"ContainerDied","Data":"ce495edb919adc1a70e8680ce556870be7891cd6340328035a4df6886ea57703"} Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.220125 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.338846 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-combined-ca-bundle\") pod \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.339042 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52ww6\" (UniqueName: \"kubernetes.io/projected/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-kube-api-access-52ww6\") pod \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.339108 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-logs\") pod \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.339389 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-logs" (OuterVolumeSpecName: "logs") pod "38c0c4b1-97f2-4bec-bba5-df9fe04663ac" (UID: "38c0c4b1-97f2-4bec-bba5-df9fe04663ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.339520 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-config-data\") pod \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\" (UID: \"38c0c4b1-97f2-4bec-bba5-df9fe04663ac\") " Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.340328 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.357941 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-kube-api-access-52ww6" (OuterVolumeSpecName: "kube-api-access-52ww6") pod "38c0c4b1-97f2-4bec-bba5-df9fe04663ac" (UID: "38c0c4b1-97f2-4bec-bba5-df9fe04663ac"). InnerVolumeSpecName "kube-api-access-52ww6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.371435 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-config-data" (OuterVolumeSpecName: "config-data") pod "38c0c4b1-97f2-4bec-bba5-df9fe04663ac" (UID: "38c0c4b1-97f2-4bec-bba5-df9fe04663ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.371960 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38c0c4b1-97f2-4bec-bba5-df9fe04663ac" (UID: "38c0c4b1-97f2-4bec-bba5-df9fe04663ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.442594 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.443130 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:34 crc kubenswrapper[4696]: I1125 10:57:34.443207 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52ww6\" (UniqueName: \"kubernetes.io/projected/38c0c4b1-97f2-4bec-bba5-df9fe04663ac-kube-api-access-52ww6\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.151821 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.152368 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38c0c4b1-97f2-4bec-bba5-df9fe04663ac","Type":"ContainerDied","Data":"a6ec50d7c19f3532e95f6da9fc6f64a7fb4f88a2f2115befaacf877ebcd5a01f"} Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.152442 4696 scope.go:117] "RemoveContainer" containerID="d9265fa05182b7a257b182f89adfc9802f8bf1f422914bbd848c5aca90cf2827" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.195591 4696 scope.go:117] "RemoveContainer" containerID="ce495edb919adc1a70e8680ce556870be7891cd6340328035a4df6886ea57703" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.203722 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.226732 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.252751 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:35 crc kubenswrapper[4696]: E1125 10:57:35.253163 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-log" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.253179 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-log" Nov 25 10:57:35 crc kubenswrapper[4696]: E1125 10:57:35.253208 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-metadata" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.253214 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-metadata" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.253398 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-log" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.253414 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" containerName="nova-metadata-metadata" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.254377 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.259003 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.259037 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.270619 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.392197 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7bgw\" (UniqueName: \"kubernetes.io/projected/769be71b-fa9b-4db6-a713-eec184e85453-kube-api-access-x7bgw\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.392284 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-config-data\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.392430 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.392464 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/769be71b-fa9b-4db6-a713-eec184e85453-logs\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.392511 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.494409 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.494479 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/769be71b-fa9b-4db6-a713-eec184e85453-logs\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.494532 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.494619 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7bgw\" (UniqueName: \"kubernetes.io/projected/769be71b-fa9b-4db6-a713-eec184e85453-kube-api-access-x7bgw\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.494936 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/769be71b-fa9b-4db6-a713-eec184e85453-logs\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.495552 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-config-data\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.501792 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.502307 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-config-data\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.511162 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.511877 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7bgw\" (UniqueName: \"kubernetes.io/projected/769be71b-fa9b-4db6-a713-eec184e85453-kube-api-access-x7bgw\") pod \"nova-metadata-0\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " pod="openstack/nova-metadata-0" Nov 25 10:57:35 crc kubenswrapper[4696]: I1125 10:57:35.589543 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:36 crc kubenswrapper[4696]: I1125 10:57:36.055483 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c0c4b1-97f2-4bec-bba5-df9fe04663ac" path="/var/lib/kubelet/pods/38c0c4b1-97f2-4bec-bba5-df9fe04663ac/volumes" Nov 25 10:57:36 crc kubenswrapper[4696]: I1125 10:57:36.098454 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:36 crc kubenswrapper[4696]: I1125 10:57:36.167528 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"769be71b-fa9b-4db6-a713-eec184e85453","Type":"ContainerStarted","Data":"ce2df37a1d324ea561aed8d03dc579f684b823885916a06937f0668d8e1c280a"} Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.182348 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"769be71b-fa9b-4db6-a713-eec184e85453","Type":"ContainerStarted","Data":"1110b7d6cb366048e5b6cdb7db426d95b23a01ca5b1650091f3044d0db44482d"} Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.182954 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"769be71b-fa9b-4db6-a713-eec184e85453","Type":"ContainerStarted","Data":"ca38991a47f9c991f8b0aa8d21dcc0aeaaa4985490a449616113279655bede26"} Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.186099 4696 generic.go:334] "Generic (PLEG): container finished" podID="ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" containerID="96e7cd861737e0e9581db9df692210348ae8c3c1c31ea2adea0c671058b1f2c5" exitCode=0 Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.186132 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfvg7" event={"ID":"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527","Type":"ContainerDied","Data":"96e7cd861737e0e9581db9df692210348ae8c3c1c31ea2adea0c671058b1f2c5"} Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.207477 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.207456337 podStartE2EDuration="2.207456337s" podCreationTimestamp="2025-11-25 10:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:37.19912656 +0000 UTC m=+1254.012743167" watchObservedRunningTime="2025-11-25 10:57:37.207456337 +0000 UTC m=+1254.021072924" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.261092 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.261367 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.409971 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.410223 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.463381 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.721258 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.815863 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.985305 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dt2q7"] Nov 25 10:57:37 crc kubenswrapper[4696]: I1125 10:57:37.985529 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerName="dnsmasq-dns" containerID="cri-o://5657b9c21cf0b28b8f20a3b0bcac6c9ab2367ea6e314788a44a66455cdccb4ae" gracePeriod=10 Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.200524 4696 generic.go:334] "Generic (PLEG): container finished" podID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerID="5657b9c21cf0b28b8f20a3b0bcac6c9ab2367ea6e314788a44a66455cdccb4ae" exitCode=0 Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.200584 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" event={"ID":"751aa451-111d-4c91-b4bc-c0d139fa14be","Type":"ContainerDied","Data":"5657b9c21cf0b28b8f20a3b0bcac6c9ab2367ea6e314788a44a66455cdccb4ae"} Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.282359 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.342852 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.342892 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.689802 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.701863 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.771516 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-sb\") pod \"751aa451-111d-4c91-b4bc-c0d139fa14be\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.771573 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-scripts\") pod \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.771655 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-svc\") pod \"751aa451-111d-4c91-b4bc-c0d139fa14be\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772388 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htp7x\" (UniqueName: \"kubernetes.io/projected/751aa451-111d-4c91-b4bc-c0d139fa14be-kube-api-access-htp7x\") pod \"751aa451-111d-4c91-b4bc-c0d139fa14be\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772464 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-nb\") pod \"751aa451-111d-4c91-b4bc-c0d139fa14be\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772486 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqf8m\" (UniqueName: \"kubernetes.io/projected/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-kube-api-access-dqf8m\") pod \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772524 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-config\") pod \"751aa451-111d-4c91-b4bc-c0d139fa14be\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772555 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-swift-storage-0\") pod \"751aa451-111d-4c91-b4bc-c0d139fa14be\" (UID: \"751aa451-111d-4c91-b4bc-c0d139fa14be\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772598 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-combined-ca-bundle\") pod \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.772618 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-config-data\") pod \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\" (UID: \"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527\") " Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.802847 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-scripts" (OuterVolumeSpecName: "scripts") pod "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" (UID: "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.802964 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-kube-api-access-dqf8m" (OuterVolumeSpecName: "kube-api-access-dqf8m") pod "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" (UID: "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527"). InnerVolumeSpecName "kube-api-access-dqf8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.803023 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751aa451-111d-4c91-b4bc-c0d139fa14be-kube-api-access-htp7x" (OuterVolumeSpecName: "kube-api-access-htp7x") pod "751aa451-111d-4c91-b4bc-c0d139fa14be" (UID: "751aa451-111d-4c91-b4bc-c0d139fa14be"). InnerVolumeSpecName "kube-api-access-htp7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.855484 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" (UID: "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.874885 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.875111 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.875551 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htp7x\" (UniqueName: \"kubernetes.io/projected/751aa451-111d-4c91-b4bc-c0d139fa14be-kube-api-access-htp7x\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.875698 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqf8m\" (UniqueName: \"kubernetes.io/projected/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-kube-api-access-dqf8m\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.888627 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "751aa451-111d-4c91-b4bc-c0d139fa14be" (UID: "751aa451-111d-4c91-b4bc-c0d139fa14be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.907102 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-config-data" (OuterVolumeSpecName: "config-data") pod "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" (UID: "ef638cd7-5ef4-4c2e-96e9-d7655bc6f527"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.918458 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-config" (OuterVolumeSpecName: "config") pod "751aa451-111d-4c91-b4bc-c0d139fa14be" (UID: "751aa451-111d-4c91-b4bc-c0d139fa14be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.924040 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "751aa451-111d-4c91-b4bc-c0d139fa14be" (UID: "751aa451-111d-4c91-b4bc-c0d139fa14be"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.932927 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "751aa451-111d-4c91-b4bc-c0d139fa14be" (UID: "751aa451-111d-4c91-b4bc-c0d139fa14be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.942832 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "751aa451-111d-4c91-b4bc-c0d139fa14be" (UID: "751aa451-111d-4c91-b4bc-c0d139fa14be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.976892 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.976918 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.976927 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.976935 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.976946 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4696]: I1125 10:57:38.976953 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751aa451-111d-4c91-b4bc-c0d139fa14be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.210847 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" event={"ID":"751aa451-111d-4c91-b4bc-c0d139fa14be","Type":"ContainerDied","Data":"b2fb44b083cd716fb074819c2eee1ae4136b225cf13bf9770753e0c9c8d23673"} Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.210906 4696 scope.go:117] "RemoveContainer" containerID="5657b9c21cf0b28b8f20a3b0bcac6c9ab2367ea6e314788a44a66455cdccb4ae" Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.210859 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dt2q7" Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.213126 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lfvg7" event={"ID":"ef638cd7-5ef4-4c2e-96e9-d7655bc6f527","Type":"ContainerDied","Data":"7be4dfae74b41fd675f56d71ff5fe142941015d7a6535cddc5db3f6ea6550c24"} Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.213157 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7be4dfae74b41fd675f56d71ff5fe142941015d7a6535cddc5db3f6ea6550c24" Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.214533 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lfvg7" Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.252192 4696 scope.go:117] "RemoveContainer" containerID="7ff9ae8b9c72802068a1f15a4ae0531f3629c49034c83f8f70e0ddfc34f3a7f7" Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.281388 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dt2q7"] Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.292883 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dt2q7"] Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.390126 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.390611 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-api" containerID="cri-o://14b71c8f64593d8205994f62562a3b8fd60a91a3c97ed9cdd437ad5ad84083b0" gracePeriod=30 Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.390399 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-log" containerID="cri-o://73fe6adabb90bd75e23a61242f35e20e563055adcaf260fcb9e3176f5e24f6df" gracePeriod=30 Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.464598 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.471841 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.472038 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-log" containerID="cri-o://ca38991a47f9c991f8b0aa8d21dcc0aeaaa4985490a449616113279655bede26" gracePeriod=30 Nov 25 10:57:39 crc kubenswrapper[4696]: I1125 10:57:39.472074 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-metadata" containerID="cri-o://1110b7d6cb366048e5b6cdb7db426d95b23a01ca5b1650091f3044d0db44482d" gracePeriod=30 Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.062328 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" path="/var/lib/kubelet/pods/751aa451-111d-4c91-b4bc-c0d139fa14be/volumes" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.256997 4696 generic.go:334] "Generic (PLEG): container finished" podID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerID="73fe6adabb90bd75e23a61242f35e20e563055adcaf260fcb9e3176f5e24f6df" exitCode=143 Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.257068 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7a7d178-8896-46b6-a2c6-b5530d20402e","Type":"ContainerDied","Data":"73fe6adabb90bd75e23a61242f35e20e563055adcaf260fcb9e3176f5e24f6df"} Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.258701 4696 generic.go:334] "Generic (PLEG): container finished" podID="769be71b-fa9b-4db6-a713-eec184e85453" containerID="1110b7d6cb366048e5b6cdb7db426d95b23a01ca5b1650091f3044d0db44482d" exitCode=0 Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.258719 4696 generic.go:334] "Generic (PLEG): container finished" podID="769be71b-fa9b-4db6-a713-eec184e85453" containerID="ca38991a47f9c991f8b0aa8d21dcc0aeaaa4985490a449616113279655bede26" exitCode=143 Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.258751 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"769be71b-fa9b-4db6-a713-eec184e85453","Type":"ContainerDied","Data":"1110b7d6cb366048e5b6cdb7db426d95b23a01ca5b1650091f3044d0db44482d"} Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.258767 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"769be71b-fa9b-4db6-a713-eec184e85453","Type":"ContainerDied","Data":"ca38991a47f9c991f8b0aa8d21dcc0aeaaa4985490a449616113279655bede26"} Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.436016 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.513267 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-config-data\") pod \"769be71b-fa9b-4db6-a713-eec184e85453\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.513325 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7bgw\" (UniqueName: \"kubernetes.io/projected/769be71b-fa9b-4db6-a713-eec184e85453-kube-api-access-x7bgw\") pod \"769be71b-fa9b-4db6-a713-eec184e85453\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.513391 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/769be71b-fa9b-4db6-a713-eec184e85453-logs\") pod \"769be71b-fa9b-4db6-a713-eec184e85453\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.513408 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-nova-metadata-tls-certs\") pod \"769be71b-fa9b-4db6-a713-eec184e85453\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.513456 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-combined-ca-bundle\") pod \"769be71b-fa9b-4db6-a713-eec184e85453\" (UID: \"769be71b-fa9b-4db6-a713-eec184e85453\") " Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.516971 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769be71b-fa9b-4db6-a713-eec184e85453-logs" (OuterVolumeSpecName: "logs") pod "769be71b-fa9b-4db6-a713-eec184e85453" (UID: "769be71b-fa9b-4db6-a713-eec184e85453"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.539068 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769be71b-fa9b-4db6-a713-eec184e85453-kube-api-access-x7bgw" (OuterVolumeSpecName: "kube-api-access-x7bgw") pod "769be71b-fa9b-4db6-a713-eec184e85453" (UID: "769be71b-fa9b-4db6-a713-eec184e85453"). InnerVolumeSpecName "kube-api-access-x7bgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.590777 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "769be71b-fa9b-4db6-a713-eec184e85453" (UID: "769be71b-fa9b-4db6-a713-eec184e85453"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.625002 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7bgw\" (UniqueName: \"kubernetes.io/projected/769be71b-fa9b-4db6-a713-eec184e85453-kube-api-access-x7bgw\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.625034 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/769be71b-fa9b-4db6-a713-eec184e85453-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.625045 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.640383 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-config-data" (OuterVolumeSpecName: "config-data") pod "769be71b-fa9b-4db6-a713-eec184e85453" (UID: "769be71b-fa9b-4db6-a713-eec184e85453"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.647174 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "769be71b-fa9b-4db6-a713-eec184e85453" (UID: "769be71b-fa9b-4db6-a713-eec184e85453"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.727136 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:40 crc kubenswrapper[4696]: I1125 10:57:40.727170 4696 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/769be71b-fa9b-4db6-a713-eec184e85453-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.032501 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.282927 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"769be71b-fa9b-4db6-a713-eec184e85453","Type":"ContainerDied","Data":"ce2df37a1d324ea561aed8d03dc579f684b823885916a06937f0668d8e1c280a"} Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.282942 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.282987 4696 scope.go:117] "RemoveContainer" containerID="1110b7d6cb366048e5b6cdb7db426d95b23a01ca5b1650091f3044d0db44482d" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.283413 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8a59b88b-c4ab-42cb-81df-8baaabbae785" containerName="nova-scheduler-scheduler" containerID="cri-o://b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3" gracePeriod=30 Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.310294 4696 scope.go:117] "RemoveContainer" containerID="ca38991a47f9c991f8b0aa8d21dcc0aeaaa4985490a449616113279655bede26" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.354040 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.361721 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.371276 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:41 crc kubenswrapper[4696]: E1125 10:57:41.371775 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerName="init" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.371794 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerName="init" Nov 25 10:57:41 crc kubenswrapper[4696]: E1125 10:57:41.371817 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-metadata" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.371824 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-metadata" Nov 25 10:57:41 crc kubenswrapper[4696]: E1125 10:57:41.371835 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" containerName="nova-manage" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.371842 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" containerName="nova-manage" Nov 25 10:57:41 crc kubenswrapper[4696]: E1125 10:57:41.371857 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerName="dnsmasq-dns" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.371863 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerName="dnsmasq-dns" Nov 25 10:57:41 crc kubenswrapper[4696]: E1125 10:57:41.371882 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-log" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.371889 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-log" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.372047 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" containerName="nova-manage" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.372058 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-metadata" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.372077 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="769be71b-fa9b-4db6-a713-eec184e85453" containerName="nova-metadata-log" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.372090 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="751aa451-111d-4c91-b4bc-c0d139fa14be" containerName="dnsmasq-dns" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.373089 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.379121 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.379124 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.382195 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.440190 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.440253 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lh5x\" (UniqueName: \"kubernetes.io/projected/b67cc31d-f86a-4851-9ac2-4936d79f8273-kube-api-access-5lh5x\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.440341 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.440371 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-config-data\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.440422 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67cc31d-f86a-4851-9ac2-4936d79f8273-logs\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.542718 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.543109 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-config-data\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.543311 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67cc31d-f86a-4851-9ac2-4936d79f8273-logs\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.543593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.543861 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lh5x\" (UniqueName: \"kubernetes.io/projected/b67cc31d-f86a-4851-9ac2-4936d79f8273-kube-api-access-5lh5x\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.543898 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67cc31d-f86a-4851-9ac2-4936d79f8273-logs\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.547067 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.547137 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-config-data\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.547287 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.565321 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lh5x\" (UniqueName: \"kubernetes.io/projected/b67cc31d-f86a-4851-9ac2-4936d79f8273-kube-api-access-5lh5x\") pod \"nova-metadata-0\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " pod="openstack/nova-metadata-0" Nov 25 10:57:41 crc kubenswrapper[4696]: I1125 10:57:41.731962 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.052657 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769be71b-fa9b-4db6-a713-eec184e85453" path="/var/lib/kubelet/pods/769be71b-fa9b-4db6-a713-eec184e85453/volumes" Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.189212 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:57:42 crc kubenswrapper[4696]: W1125 10:57:42.192811 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb67cc31d_f86a_4851_9ac2_4936d79f8273.slice/crio-6d0e242c0c170f76d0644b9a7c308231db3c748fb05960f21518d21c58b19df6 WatchSource:0}: Error finding container 6d0e242c0c170f76d0644b9a7c308231db3c748fb05960f21518d21c58b19df6: Status 404 returned error can't find the container with id 6d0e242c0c170f76d0644b9a7c308231db3c748fb05960f21518d21c58b19df6 Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.292986 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b67cc31d-f86a-4851-9ac2-4936d79f8273","Type":"ContainerStarted","Data":"6d0e242c0c170f76d0644b9a7c308231db3c748fb05960f21518d21c58b19df6"} Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.297392 4696 generic.go:334] "Generic (PLEG): container finished" podID="4b10ca5a-c03c-407e-9d19-c649df1f9711" containerID="4d316d2a2a1f34711da8f0a58a99152d05248979396e27f75374e53e5f17b288" exitCode=0 Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.297431 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfdng" event={"ID":"4b10ca5a-c03c-407e-9d19-c649df1f9711","Type":"ContainerDied","Data":"4d316d2a2a1f34711da8f0a58a99152d05248979396e27f75374e53e5f17b288"} Nov 25 10:57:42 crc kubenswrapper[4696]: E1125 10:57:42.412981 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 10:57:42 crc kubenswrapper[4696]: E1125 10:57:42.415001 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 10:57:42 crc kubenswrapper[4696]: E1125 10:57:42.416516 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 10:57:42 crc kubenswrapper[4696]: E1125 10:57:42.416552 4696 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8a59b88b-c4ab-42cb-81df-8baaabbae785" containerName="nova-scheduler-scheduler" Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.761654 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c8fd89c8d-5vb5f" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Nov 25 10:57:42 crc kubenswrapper[4696]: I1125 10:57:42.761788 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.307352 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b67cc31d-f86a-4851-9ac2-4936d79f8273","Type":"ContainerStarted","Data":"a1e21147c6240a548d496dccb577f9e853613f365e7d7265b9d79e18445e106e"} Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.307553 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b67cc31d-f86a-4851-9ac2-4936d79f8273","Type":"ContainerStarted","Data":"318d4f95c617c7edcb581ef02df1cde2006040322c6877f5f5babd4e33a4ef63"} Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.339248 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.339225032 podStartE2EDuration="2.339225032s" podCreationTimestamp="2025-11-25 10:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:43.33415458 +0000 UTC m=+1260.147771167" watchObservedRunningTime="2025-11-25 10:57:43.339225032 +0000 UTC m=+1260.152841619" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.705863 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.782160 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bdkq\" (UniqueName: \"kubernetes.io/projected/4b10ca5a-c03c-407e-9d19-c649df1f9711-kube-api-access-2bdkq\") pod \"4b10ca5a-c03c-407e-9d19-c649df1f9711\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.782217 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-scripts\") pod \"4b10ca5a-c03c-407e-9d19-c649df1f9711\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.783097 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-config-data\") pod \"4b10ca5a-c03c-407e-9d19-c649df1f9711\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.783473 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-combined-ca-bundle\") pod \"4b10ca5a-c03c-407e-9d19-c649df1f9711\" (UID: \"4b10ca5a-c03c-407e-9d19-c649df1f9711\") " Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.795047 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b10ca5a-c03c-407e-9d19-c649df1f9711-kube-api-access-2bdkq" (OuterVolumeSpecName: "kube-api-access-2bdkq") pod "4b10ca5a-c03c-407e-9d19-c649df1f9711" (UID: "4b10ca5a-c03c-407e-9d19-c649df1f9711"). InnerVolumeSpecName "kube-api-access-2bdkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.803303 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-scripts" (OuterVolumeSpecName: "scripts") pod "4b10ca5a-c03c-407e-9d19-c649df1f9711" (UID: "4b10ca5a-c03c-407e-9d19-c649df1f9711"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.810873 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b10ca5a-c03c-407e-9d19-c649df1f9711" (UID: "4b10ca5a-c03c-407e-9d19-c649df1f9711"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.813818 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-config-data" (OuterVolumeSpecName: "config-data") pod "4b10ca5a-c03c-407e-9d19-c649df1f9711" (UID: "4b10ca5a-c03c-407e-9d19-c649df1f9711"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.885812 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.886060 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.886136 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bdkq\" (UniqueName: \"kubernetes.io/projected/4b10ca5a-c03c-407e-9d19-c649df1f9711-kube-api-access-2bdkq\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:43 crc kubenswrapper[4696]: I1125 10:57:43.886193 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b10ca5a-c03c-407e-9d19-c649df1f9711-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.335102 4696 generic.go:334] "Generic (PLEG): container finished" podID="8a59b88b-c4ab-42cb-81df-8baaabbae785" containerID="b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3" exitCode=0 Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.335195 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a59b88b-c4ab-42cb-81df-8baaabbae785","Type":"ContainerDied","Data":"b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3"} Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.349782 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xfdng" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.350183 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xfdng" event={"ID":"4b10ca5a-c03c-407e-9d19-c649df1f9711","Type":"ContainerDied","Data":"a6da4be66c9513f840d851b033368a92d9f41e7786c30b217578f64f8a9b5257"} Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.350201 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6da4be66c9513f840d851b033368a92d9f41e7786c30b217578f64f8a9b5257" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.439413 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:57:44 crc kubenswrapper[4696]: E1125 10:57:44.440089 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b10ca5a-c03c-407e-9d19-c649df1f9711" containerName="nova-cell1-conductor-db-sync" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.440107 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b10ca5a-c03c-407e-9d19-c649df1f9711" containerName="nova-cell1-conductor-db-sync" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.440286 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b10ca5a-c03c-407e-9d19-c649df1f9711" containerName="nova-cell1-conductor-db-sync" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.440917 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.443626 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.458327 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.505436 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984434ed-24d1-4090-a7d9-c38b39617b74-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.506944 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984434ed-24d1-4090-a7d9-c38b39617b74-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.507013 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7p4q\" (UniqueName: \"kubernetes.io/projected/984434ed-24d1-4090-a7d9-c38b39617b74-kube-api-access-q7p4q\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.609237 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984434ed-24d1-4090-a7d9-c38b39617b74-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.609414 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984434ed-24d1-4090-a7d9-c38b39617b74-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.609452 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7p4q\" (UniqueName: \"kubernetes.io/projected/984434ed-24d1-4090-a7d9-c38b39617b74-kube-api-access-q7p4q\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.615110 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984434ed-24d1-4090-a7d9-c38b39617b74-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.615571 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984434ed-24d1-4090-a7d9-c38b39617b74-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.627768 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7p4q\" (UniqueName: \"kubernetes.io/projected/984434ed-24d1-4090-a7d9-c38b39617b74-kube-api-access-q7p4q\") pod \"nova-cell1-conductor-0\" (UID: \"984434ed-24d1-4090-a7d9-c38b39617b74\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.759829 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:44 crc kubenswrapper[4696]: I1125 10:57:44.900178 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.023563 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pntt4\" (UniqueName: \"kubernetes.io/projected/8a59b88b-c4ab-42cb-81df-8baaabbae785-kube-api-access-pntt4\") pod \"8a59b88b-c4ab-42cb-81df-8baaabbae785\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.023798 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-config-data\") pod \"8a59b88b-c4ab-42cb-81df-8baaabbae785\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.023840 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-combined-ca-bundle\") pod \"8a59b88b-c4ab-42cb-81df-8baaabbae785\" (UID: \"8a59b88b-c4ab-42cb-81df-8baaabbae785\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.039009 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a59b88b-c4ab-42cb-81df-8baaabbae785-kube-api-access-pntt4" (OuterVolumeSpecName: "kube-api-access-pntt4") pod "8a59b88b-c4ab-42cb-81df-8baaabbae785" (UID: "8a59b88b-c4ab-42cb-81df-8baaabbae785"). InnerVolumeSpecName "kube-api-access-pntt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.065621 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a59b88b-c4ab-42cb-81df-8baaabbae785" (UID: "8a59b88b-c4ab-42cb-81df-8baaabbae785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.069875 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-config-data" (OuterVolumeSpecName: "config-data") pod "8a59b88b-c4ab-42cb-81df-8baaabbae785" (UID: "8a59b88b-c4ab-42cb-81df-8baaabbae785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.125676 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.125706 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a59b88b-c4ab-42cb-81df-8baaabbae785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.125716 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pntt4\" (UniqueName: \"kubernetes.io/projected/8a59b88b-c4ab-42cb-81df-8baaabbae785-kube-api-access-pntt4\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.321924 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.358247 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"984434ed-24d1-4090-a7d9-c38b39617b74","Type":"ContainerStarted","Data":"618837d46781415619a8d1f2262e12826445236e403d1844869da9213577847f"} Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.360272 4696 generic.go:334] "Generic (PLEG): container finished" podID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerID="14b71c8f64593d8205994f62562a3b8fd60a91a3c97ed9cdd437ad5ad84083b0" exitCode=0 Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.360330 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7a7d178-8896-46b6-a2c6-b5530d20402e","Type":"ContainerDied","Data":"14b71c8f64593d8205994f62562a3b8fd60a91a3c97ed9cdd437ad5ad84083b0"} Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.362052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a59b88b-c4ab-42cb-81df-8baaabbae785","Type":"ContainerDied","Data":"9bea28ce411633618cf456894816714d588200a056751677f307a25d48e09abf"} Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.362111 4696 scope.go:117] "RemoveContainer" containerID="b8480d2e73b3841cd38f86c01a3f48761ead06ebf12e7fccff0cc42a79c2c2e3" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.362172 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.403229 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.415934 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.424258 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.432177 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:45 crc kubenswrapper[4696]: E1125 10:57:45.435021 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a59b88b-c4ab-42cb-81df-8baaabbae785" containerName="nova-scheduler-scheduler" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.435067 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a59b88b-c4ab-42cb-81df-8baaabbae785" containerName="nova-scheduler-scheduler" Nov 25 10:57:45 crc kubenswrapper[4696]: E1125 10:57:45.435091 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-api" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.435102 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-api" Nov 25 10:57:45 crc kubenswrapper[4696]: E1125 10:57:45.435123 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-log" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.435130 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-log" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.435442 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a59b88b-c4ab-42cb-81df-8baaabbae785" containerName="nova-scheduler-scheduler" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.435470 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-log" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.435484 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" containerName="nova-api-api" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.436199 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.438427 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.444890 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.542529 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a7d178-8896-46b6-a2c6-b5530d20402e-logs\") pod \"e7a7d178-8896-46b6-a2c6-b5530d20402e\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.542643 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-config-data\") pod \"e7a7d178-8896-46b6-a2c6-b5530d20402e\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.542691 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-combined-ca-bundle\") pod \"e7a7d178-8896-46b6-a2c6-b5530d20402e\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.542765 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlqv2\" (UniqueName: \"kubernetes.io/projected/e7a7d178-8896-46b6-a2c6-b5530d20402e-kube-api-access-rlqv2\") pod \"e7a7d178-8896-46b6-a2c6-b5530d20402e\" (UID: \"e7a7d178-8896-46b6-a2c6-b5530d20402e\") " Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.542989 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.543020 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lchpx\" (UniqueName: \"kubernetes.io/projected/c6f46bac-b868-4767-8efb-5fdd89267c66-kube-api-access-lchpx\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.543064 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-config-data\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.543081 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a7d178-8896-46b6-a2c6-b5530d20402e-logs" (OuterVolumeSpecName: "logs") pod "e7a7d178-8896-46b6-a2c6-b5530d20402e" (UID: "e7a7d178-8896-46b6-a2c6-b5530d20402e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.543244 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7a7d178-8896-46b6-a2c6-b5530d20402e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.546384 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a7d178-8896-46b6-a2c6-b5530d20402e-kube-api-access-rlqv2" (OuterVolumeSpecName: "kube-api-access-rlqv2") pod "e7a7d178-8896-46b6-a2c6-b5530d20402e" (UID: "e7a7d178-8896-46b6-a2c6-b5530d20402e"). InnerVolumeSpecName "kube-api-access-rlqv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.567757 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7a7d178-8896-46b6-a2c6-b5530d20402e" (UID: "e7a7d178-8896-46b6-a2c6-b5530d20402e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.574376 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-config-data" (OuterVolumeSpecName: "config-data") pod "e7a7d178-8896-46b6-a2c6-b5530d20402e" (UID: "e7a7d178-8896-46b6-a2c6-b5530d20402e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.626959 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.627195 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="87ef4570-7d70-44f8-9249-62d32b384642" containerName="kube-state-metrics" containerID="cri-o://8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9" gracePeriod=30 Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.644857 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-config-data\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.645029 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.645058 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lchpx\" (UniqueName: \"kubernetes.io/projected/c6f46bac-b868-4767-8efb-5fdd89267c66-kube-api-access-lchpx\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.645539 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.645560 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7a7d178-8896-46b6-a2c6-b5530d20402e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.645572 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlqv2\" (UniqueName: \"kubernetes.io/projected/e7a7d178-8896-46b6-a2c6-b5530d20402e-kube-api-access-rlqv2\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.648533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-config-data\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.648583 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.665747 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lchpx\" (UniqueName: \"kubernetes.io/projected/c6f46bac-b868-4767-8efb-5fdd89267c66-kube-api-access-lchpx\") pod \"nova-scheduler-0\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " pod="openstack/nova-scheduler-0" Nov 25 10:57:45 crc kubenswrapper[4696]: I1125 10:57:45.776304 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.058657 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a59b88b-c4ab-42cb-81df-8baaabbae785" path="/var/lib/kubelet/pods/8a59b88b-c4ab-42cb-81df-8baaabbae785/volumes" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.105117 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.157003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-577xl\" (UniqueName: \"kubernetes.io/projected/87ef4570-7d70-44f8-9249-62d32b384642-kube-api-access-577xl\") pod \"87ef4570-7d70-44f8-9249-62d32b384642\" (UID: \"87ef4570-7d70-44f8-9249-62d32b384642\") " Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.168490 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ef4570-7d70-44f8-9249-62d32b384642-kube-api-access-577xl" (OuterVolumeSpecName: "kube-api-access-577xl") pod "87ef4570-7d70-44f8-9249-62d32b384642" (UID: "87ef4570-7d70-44f8-9249-62d32b384642"). InnerVolumeSpecName "kube-api-access-577xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.259615 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-577xl\" (UniqueName: \"kubernetes.io/projected/87ef4570-7d70-44f8-9249-62d32b384642-kube-api-access-577xl\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.356586 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.372455 4696 generic.go:334] "Generic (PLEG): container finished" podID="87ef4570-7d70-44f8-9249-62d32b384642" containerID="8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9" exitCode=2 Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.372525 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"87ef4570-7d70-44f8-9249-62d32b384642","Type":"ContainerDied","Data":"8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9"} Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.372540 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.372562 4696 scope.go:117] "RemoveContainer" containerID="8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.372550 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"87ef4570-7d70-44f8-9249-62d32b384642","Type":"ContainerDied","Data":"7fda7bf0de2bcc7e7449a0d17f427ff3034d4e190d4ece264a28cf631af9d6da"} Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.374164 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6f46bac-b868-4767-8efb-5fdd89267c66","Type":"ContainerStarted","Data":"65e89b9664d83b70315f20ff40b3b363a7a1e4bf87f05b6e87a3b80b7285bb51"} Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.376108 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"984434ed-24d1-4090-a7d9-c38b39617b74","Type":"ContainerStarted","Data":"1789ffbb047675844be9226967105fffc664438a146d3e1e3738c6c4f3d56296"} Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.377007 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.389126 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7a7d178-8896-46b6-a2c6-b5530d20402e","Type":"ContainerDied","Data":"950f19bb5acfceef768a030d4223ba6081b4c59a2fe39d6981d068642328586d"} Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.389220 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.402261 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.4022459769999998 podStartE2EDuration="2.402245977s" podCreationTimestamp="2025-11-25 10:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:46.396249939 +0000 UTC m=+1263.209866526" watchObservedRunningTime="2025-11-25 10:57:46.402245977 +0000 UTC m=+1263.215862564" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.423533 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.432249 4696 scope.go:117] "RemoveContainer" containerID="8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9" Nov 25 10:57:46 crc kubenswrapper[4696]: E1125 10:57:46.432777 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9\": container with ID starting with 8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9 not found: ID does not exist" containerID="8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.432810 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9"} err="failed to get container status \"8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9\": rpc error: code = NotFound desc = could not find container \"8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9\": container with ID starting with 8106fa505a31360d66a7c6a129fe78748b116ac73a0550c2091aeedc332e02b9 not found: ID does not exist" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.432830 4696 scope.go:117] "RemoveContainer" containerID="14b71c8f64593d8205994f62562a3b8fd60a91a3c97ed9cdd437ad5ad84083b0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.434245 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.459889 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.469837 4696 scope.go:117] "RemoveContainer" containerID="73fe6adabb90bd75e23a61242f35e20e563055adcaf260fcb9e3176f5e24f6df" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.503688 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.516453 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: E1125 10:57:46.517123 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ef4570-7d70-44f8-9249-62d32b384642" containerName="kube-state-metrics" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.517146 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ef4570-7d70-44f8-9249-62d32b384642" containerName="kube-state-metrics" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.517402 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ef4570-7d70-44f8-9249-62d32b384642" containerName="kube-state-metrics" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.518116 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.520497 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.520634 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.528709 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.530281 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.532031 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.547605 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.557047 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.563996 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplvm\" (UniqueName: \"kubernetes.io/projected/9ed1a610-3fb4-4643-a481-724f5036b836-kube-api-access-gplvm\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.564084 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.564105 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.564176 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666267 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-config-data\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666652 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666696 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql8mz\" (UniqueName: \"kubernetes.io/projected/3041af3a-3764-4e9d-aedf-c45ef8e34edf-kube-api-access-ql8mz\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666779 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666853 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplvm\" (UniqueName: \"kubernetes.io/projected/9ed1a610-3fb4-4643-a481-724f5036b836-kube-api-access-gplvm\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.666991 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041af3a-3764-4e9d-aedf-c45ef8e34edf-logs\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.671313 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.672403 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.675718 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed1a610-3fb4-4643-a481-724f5036b836-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.686543 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplvm\" (UniqueName: \"kubernetes.io/projected/9ed1a610-3fb4-4643-a481-724f5036b836-kube-api-access-gplvm\") pod \"kube-state-metrics-0\" (UID: \"9ed1a610-3fb4-4643-a481-724f5036b836\") " pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.734008 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.734339 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.769140 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.769273 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041af3a-3764-4e9d-aedf-c45ef8e34edf-logs\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.769319 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-config-data\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.769345 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql8mz\" (UniqueName: \"kubernetes.io/projected/3041af3a-3764-4e9d-aedf-c45ef8e34edf-kube-api-access-ql8mz\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.779228 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041af3a-3764-4e9d-aedf-c45ef8e34edf-logs\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.789097 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-config-data\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.793275 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.793452 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql8mz\" (UniqueName: \"kubernetes.io/projected/3041af3a-3764-4e9d-aedf-c45ef8e34edf-kube-api-access-ql8mz\") pod \"nova-api-0\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " pod="openstack/nova-api-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.849041 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:57:46 crc kubenswrapper[4696]: I1125 10:57:46.855848 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:57:47 crc kubenswrapper[4696]: I1125 10:57:47.372234 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:57:47 crc kubenswrapper[4696]: I1125 10:57:47.411979 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6f46bac-b868-4767-8efb-5fdd89267c66","Type":"ContainerStarted","Data":"e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1"} Nov 25 10:57:47 crc kubenswrapper[4696]: I1125 10:57:47.412979 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9ed1a610-3fb4-4643-a481-724f5036b836","Type":"ContainerStarted","Data":"6d01b058bda0853af85b6e6f9801a6766b66cbd1a4de24247d3c2da8fef9a5f5"} Nov 25 10:57:47 crc kubenswrapper[4696]: I1125 10:57:47.448365 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.44834846 podStartE2EDuration="2.44834846s" podCreationTimestamp="2025-11-25 10:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:47.447246228 +0000 UTC m=+1264.260862815" watchObservedRunningTime="2025-11-25 10:57:47.44834846 +0000 UTC m=+1264.261965047" Nov 25 10:57:47 crc kubenswrapper[4696]: I1125 10:57:47.473732 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.055080 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ef4570-7d70-44f8-9249-62d32b384642" path="/var/lib/kubelet/pods/87ef4570-7d70-44f8-9249-62d32b384642/volumes" Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.365104 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a7d178-8896-46b6-a2c6-b5530d20402e" path="/var/lib/kubelet/pods/e7a7d178-8896-46b6-a2c6-b5530d20402e/volumes" Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.432119 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9ed1a610-3fb4-4643-a481-724f5036b836","Type":"ContainerStarted","Data":"87526e7b4db97177f0a8baf6506f952fd862c89aef98d57cffe8d1005ccdc463"} Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.433235 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.438113 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3041af3a-3764-4e9d-aedf-c45ef8e34edf","Type":"ContainerStarted","Data":"71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc"} Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.438156 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3041af3a-3764-4e9d-aedf-c45ef8e34edf","Type":"ContainerStarted","Data":"2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a"} Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.438167 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3041af3a-3764-4e9d-aedf-c45ef8e34edf","Type":"ContainerStarted","Data":"f0932dcb782465230d1609f5ea390a9e8e0086ff5b4e8cbe210974575faf723c"} Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.438586 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.438835 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-central-agent" containerID="cri-o://47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5" gracePeriod=30 Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.438998 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-notification-agent" containerID="cri-o://e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f" gracePeriod=30 Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.439061 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="sg-core" containerID="cri-o://20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71" gracePeriod=30 Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.439453 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="proxy-httpd" containerID="cri-o://63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487" gracePeriod=30 Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.464216 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.098757415 podStartE2EDuration="2.464201026s" podCreationTimestamp="2025-11-25 10:57:46 +0000 UTC" firstStartedPulling="2025-11-25 10:57:47.378104634 +0000 UTC m=+1264.191721221" lastFinishedPulling="2025-11-25 10:57:47.743548245 +0000 UTC m=+1264.557164832" observedRunningTime="2025-11-25 10:57:48.463707847 +0000 UTC m=+1265.277324434" watchObservedRunningTime="2025-11-25 10:57:48.464201026 +0000 UTC m=+1265.277817613" Nov 25 10:57:48 crc kubenswrapper[4696]: I1125 10:57:48.503092 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5030724109999998 podStartE2EDuration="2.503072411s" podCreationTimestamp="2025-11-25 10:57:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:48.487929287 +0000 UTC m=+1265.301545874" watchObservedRunningTime="2025-11-25 10:57:48.503072411 +0000 UTC m=+1265.316688998" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.166855 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.287762 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h88gp\" (UniqueName: \"kubernetes.io/projected/28158e14-5724-410c-9d4c-5ca5563f6f87-kube-api-access-h88gp\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.287821 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-config-data\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.287846 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-scripts\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.287904 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-secret-key\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.287924 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28158e14-5724-410c-9d4c-5ca5563f6f87-logs\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.287949 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-combined-ca-bundle\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.288011 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-tls-certs\") pod \"28158e14-5724-410c-9d4c-5ca5563f6f87\" (UID: \"28158e14-5724-410c-9d4c-5ca5563f6f87\") " Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.288570 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28158e14-5724-410c-9d4c-5ca5563f6f87-logs" (OuterVolumeSpecName: "logs") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.294907 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28158e14-5724-410c-9d4c-5ca5563f6f87-kube-api-access-h88gp" (OuterVolumeSpecName: "kube-api-access-h88gp") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "kube-api-access-h88gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.302395 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.330258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.338210 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-scripts" (OuterVolumeSpecName: "scripts") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.349537 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-config-data" (OuterVolumeSpecName: "config-data") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.370588 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "28158e14-5724-410c-9d4c-5ca5563f6f87" (UID: "28158e14-5724-410c-9d4c-5ca5563f6f87"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.389847 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h88gp\" (UniqueName: \"kubernetes.io/projected/28158e14-5724-410c-9d4c-5ca5563f6f87-kube-api-access-h88gp\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.390092 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.390211 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28158e14-5724-410c-9d4c-5ca5563f6f87-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.390302 4696 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.390386 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28158e14-5724-410c-9d4c-5ca5563f6f87-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.390467 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.390547 4696 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28158e14-5724-410c-9d4c-5ca5563f6f87-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.447501 4696 generic.go:334] "Generic (PLEG): container finished" podID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerID="63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487" exitCode=0 Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.447560 4696 generic.go:334] "Generic (PLEG): container finished" podID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerID="20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71" exitCode=2 Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.447568 4696 generic.go:334] "Generic (PLEG): container finished" podID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerID="47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5" exitCode=0 Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.448350 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerDied","Data":"63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487"} Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.448477 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerDied","Data":"20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71"} Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.448761 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerDied","Data":"47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5"} Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.449267 4696 generic.go:334] "Generic (PLEG): container finished" podID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerID="2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154" exitCode=137 Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.449367 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerDied","Data":"2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154"} Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.449438 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8fd89c8d-5vb5f" event={"ID":"28158e14-5724-410c-9d4c-5ca5563f6f87","Type":"ContainerDied","Data":"2e7adca82ed91f7e8f4bf14ec40ea94cb01cac7b98d14a1bc9e48e4198bd2743"} Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.449499 4696 scope.go:117] "RemoveContainer" containerID="91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.449654 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8fd89c8d-5vb5f" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.488728 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c8fd89c8d-5vb5f"] Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.500232 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c8fd89c8d-5vb5f"] Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.632581 4696 scope.go:117] "RemoveContainer" containerID="2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.700625 4696 scope.go:117] "RemoveContainer" containerID="91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829" Nov 25 10:57:49 crc kubenswrapper[4696]: E1125 10:57:49.702357 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829\": container with ID starting with 91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829 not found: ID does not exist" containerID="91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.702407 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829"} err="failed to get container status \"91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829\": rpc error: code = NotFound desc = could not find container \"91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829\": container with ID starting with 91761e996fe329220032372dffdc03352be2ef1645328834ab04dc4b5ac0d829 not found: ID does not exist" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.702439 4696 scope.go:117] "RemoveContainer" containerID="2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154" Nov 25 10:57:49 crc kubenswrapper[4696]: E1125 10:57:49.702825 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154\": container with ID starting with 2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154 not found: ID does not exist" containerID="2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154" Nov 25 10:57:49 crc kubenswrapper[4696]: I1125 10:57:49.702873 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154"} err="failed to get container status \"2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154\": rpc error: code = NotFound desc = could not find container \"2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154\": container with ID starting with 2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154 not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.059530 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" path="/var/lib/kubelet/pods/28158e14-5724-410c-9d4c-5ca5563f6f87/volumes" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.181905 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309250 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-log-httpd\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309294 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-scripts\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309430 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h8x7\" (UniqueName: \"kubernetes.io/projected/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-kube-api-access-7h8x7\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309450 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309500 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-config-data\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309551 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-sg-core-conf-yaml\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.309578 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-run-httpd\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.310105 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.310064 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.318913 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-scripts" (OuterVolumeSpecName: "scripts") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.323440 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-kube-api-access-7h8x7" (OuterVolumeSpecName: "kube-api-access-7h8x7") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "kube-api-access-7h8x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.356254 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.410709 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.410844 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle\") pod \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\" (UID: \"8088aa1a-ff77-4c8c-991f-1ee40f7613a9\") " Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.411467 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.411492 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.411503 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.411516 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.411529 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h8x7\" (UniqueName: \"kubernetes.io/projected/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-kube-api-access-7h8x7\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: W1125 10:57:50.411609 4696 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8088aa1a-ff77-4c8c-991f-1ee40f7613a9/volumes/kubernetes.io~secret/combined-ca-bundle Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.411622 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.447369 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-config-data" (OuterVolumeSpecName: "config-data") pod "8088aa1a-ff77-4c8c-991f-1ee40f7613a9" (UID: "8088aa1a-ff77-4c8c-991f-1ee40f7613a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.461694 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerDied","Data":"e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f"} Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.461782 4696 scope.go:117] "RemoveContainer" containerID="63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.461626 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.462138 4696 generic.go:334] "Generic (PLEG): container finished" podID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerID="e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f" exitCode=0 Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.462211 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8088aa1a-ff77-4c8c-991f-1ee40f7613a9","Type":"ContainerDied","Data":"81de1837540ebb8ce8b2c8e9d68bd74e558276860c61e7146e99ef9862bb06aa"} Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.495734 4696 scope.go:117] "RemoveContainer" containerID="20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.507805 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.513396 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.513431 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8088aa1a-ff77-4c8c-991f-1ee40f7613a9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.518043 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.526414 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529810 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="proxy-httpd" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529842 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="proxy-httpd" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529855 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon-log" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529862 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon-log" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529876 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="sg-core" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529882 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="sg-core" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529897 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-central-agent" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529903 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-central-agent" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529926 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529932 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529947 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529954 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.529966 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-notification-agent" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.529972 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-notification-agent" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530188 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530202 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="sg-core" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530209 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530232 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="28158e14-5724-410c-9d4c-5ca5563f6f87" containerName="horizon-log" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530242 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="proxy-httpd" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530249 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-notification-agent" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.530260 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" containerName="ceilometer-central-agent" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.531904 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.543243 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.543907 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.544535 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.558150 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.558679 4696 scope.go:117] "RemoveContainer" containerID="e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.612426 4696 scope.go:117] "RemoveContainer" containerID="47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.684037 4696 scope.go:117] "RemoveContainer" containerID="63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.684550 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487\": container with ID starting with 63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487 not found: ID does not exist" containerID="63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.684592 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487"} err="failed to get container status \"63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487\": rpc error: code = NotFound desc = could not find container \"63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487\": container with ID starting with 63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487 not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.684622 4696 scope.go:117] "RemoveContainer" containerID="20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.685030 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71\": container with ID starting with 20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71 not found: ID does not exist" containerID="20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.685065 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71"} err="failed to get container status \"20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71\": rpc error: code = NotFound desc = could not find container \"20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71\": container with ID starting with 20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71 not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.685085 4696 scope.go:117] "RemoveContainer" containerID="e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.685355 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f\": container with ID starting with e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f not found: ID does not exist" containerID="e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.685390 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f"} err="failed to get container status \"e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f\": rpc error: code = NotFound desc = could not find container \"e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f\": container with ID starting with e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.685410 4696 scope.go:117] "RemoveContainer" containerID="47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5" Nov 25 10:57:50 crc kubenswrapper[4696]: E1125 10:57:50.685639 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5\": container with ID starting with 47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5 not found: ID does not exist" containerID="47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.685656 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5"} err="failed to get container status \"47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5\": rpc error: code = NotFound desc = could not find container \"47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5\": container with ID starting with 47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5 not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.720644 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.720763 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-run-httpd\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.720880 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.720972 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.721062 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-scripts\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.721121 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zm6r\" (UniqueName: \"kubernetes.io/projected/98853f51-ccc5-40df-8f01-00f2fabe8668-kube-api-access-5zm6r\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.721211 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-config-data\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.721237 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-log-httpd\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.777409 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823044 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823170 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-scripts\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823202 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zm6r\" (UniqueName: \"kubernetes.io/projected/98853f51-ccc5-40df-8f01-00f2fabe8668-kube-api-access-5zm6r\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823243 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-config-data\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823262 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-log-httpd\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823298 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823316 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-run-httpd\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.823936 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-run-httpd\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.824037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-log-httpd\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.830931 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-scripts\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.831191 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.832063 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-config-data\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.838632 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.842097 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.851172 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zm6r\" (UniqueName: \"kubernetes.io/projected/98853f51-ccc5-40df-8f01-00f2fabe8668-kube-api-access-5zm6r\") pod \"ceilometer-0\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " pod="openstack/ceilometer-0" Nov 25 10:57:50 crc kubenswrapper[4696]: I1125 10:57:50.877286 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:57:51 crc kubenswrapper[4696]: I1125 10:57:51.358057 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:57:51 crc kubenswrapper[4696]: I1125 10:57:51.477215 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerStarted","Data":"1831648904a1d1122bf85ce2c4321836792ec5239d1647b95f750be7cee2c92b"} Nov 25 10:57:51 crc kubenswrapper[4696]: I1125 10:57:51.732697 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:57:51 crc kubenswrapper[4696]: I1125 10:57:51.732745 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:57:52 crc kubenswrapper[4696]: I1125 10:57:52.053656 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8088aa1a-ff77-4c8c-991f-1ee40f7613a9" path="/var/lib/kubelet/pods/8088aa1a-ff77-4c8c-991f-1ee40f7613a9/volumes" Nov 25 10:57:52 crc kubenswrapper[4696]: I1125 10:57:52.492737 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerStarted","Data":"0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb"} Nov 25 10:57:52 crc kubenswrapper[4696]: I1125 10:57:52.745834 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:57:52 crc kubenswrapper[4696]: I1125 10:57:52.746364 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:57:53 crc kubenswrapper[4696]: I1125 10:57:53.503225 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerStarted","Data":"659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb"} Nov 25 10:57:53 crc kubenswrapper[4696]: I1125 10:57:53.503532 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerStarted","Data":"f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177"} Nov 25 10:57:54 crc kubenswrapper[4696]: I1125 10:57:54.785105 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 10:57:55 crc kubenswrapper[4696]: I1125 10:57:55.546893 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerStarted","Data":"f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330"} Nov 25 10:57:55 crc kubenswrapper[4696]: I1125 10:57:55.547504 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:57:55 crc kubenswrapper[4696]: I1125 10:57:55.576776 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.501991276 podStartE2EDuration="5.576749116s" podCreationTimestamp="2025-11-25 10:57:50 +0000 UTC" firstStartedPulling="2025-11-25 10:57:51.360921388 +0000 UTC m=+1268.174537975" lastFinishedPulling="2025-11-25 10:57:54.435679228 +0000 UTC m=+1271.249295815" observedRunningTime="2025-11-25 10:57:55.568591686 +0000 UTC m=+1272.382208273" watchObservedRunningTime="2025-11-25 10:57:55.576749116 +0000 UTC m=+1272.390365713" Nov 25 10:57:55 crc kubenswrapper[4696]: I1125 10:57:55.777182 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:57:55 crc kubenswrapper[4696]: I1125 10:57:55.822371 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:57:56 crc kubenswrapper[4696]: I1125 10:57:56.593441 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:57:56 crc kubenswrapper[4696]: I1125 10:57:56.856486 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:57:56 crc kubenswrapper[4696]: I1125 10:57:56.856545 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:57:56 crc kubenswrapper[4696]: I1125 10:57:56.863483 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:57:57 crc kubenswrapper[4696]: I1125 10:57:57.941150 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:57:57 crc kubenswrapper[4696]: I1125 10:57:57.941269 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:58:00 crc kubenswrapper[4696]: I1125 10:58:00.801656 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:58:00 crc kubenswrapper[4696]: I1125 10:58:00.802133 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:58:00 crc kubenswrapper[4696]: I1125 10:58:00.802202 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 10:58:00 crc kubenswrapper[4696]: I1125 10:58:00.803401 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:58:00 crc kubenswrapper[4696]: I1125 10:58:00.803513 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769" gracePeriod=600 Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.613177 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769" exitCode=0 Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.613241 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769"} Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.613299 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"8a6788bdae639f3c35f45af42113cd2f9e953a8da542ed0cc824a9b88a8dd56b"} Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.613332 4696 scope.go:117] "RemoveContainer" containerID="33cc4b473fa3af53c05eee091add2db7db56314c26dffae135812139c5ec3acf" Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.742598 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.744087 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:58:01 crc kubenswrapper[4696]: I1125 10:58:01.749830 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:58:02 crc kubenswrapper[4696]: I1125 10:58:02.643282 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:58:03 crc kubenswrapper[4696]: E1125 10:58:03.260320 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-conmon-47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-conmon-e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-conmon-20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28158e14_5724_410c_9d4c_5ca5563f6f87.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-conmon-63a703ef2612629f4b3a131f64b1b5481844905b6e39b9d9fef10f22f2291487.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-47e0476dbb9d157777119e06d09cc65045e3ddd401012f22e291d8a955458ce5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-20a8a06a3acb8c208154b58d17104ba8e963f0ac9eee128c37a4505bfa741d71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac9983ff_4a9b_456b_8a9c_7e5857aad5d2.slice/crio-b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28158e14_5724_410c_9d4c_5ca5563f6f87.slice/crio-conmon-2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28158e14_5724_410c_9d4c_5ca5563f6f87.slice/crio-2e7adca82ed91f7e8f4bf14ec40ea94cb01cac7b98d14a1bc9e48e4198bd2743\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28158e14_5724_410c_9d4c_5ca5563f6f87.slice/crio-2bbaddbf18e5cd5b9837ac4d01e03100b0b544df90d12a3aad0a64d6d7388154.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23fb3f29_8c86_4ba7_8260_4ead78279db6.slice/crio-3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-e765ce65481c47ceb5a5be22dae920fa03c18379c12e73e30c25c59b997f214f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8088aa1a_ff77_4c8c_991f_1ee40f7613a9.slice/crio-81de1837540ebb8ce8b2c8e9d68bd74e558276860c61e7146e99ef9862bb06aa\": RecentStats: unable to find data in memory cache]" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.387137 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.569909 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-combined-ca-bundle\") pod \"23fb3f29-8c86-4ba7-8260-4ead78279db6\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.570033 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-config-data\") pod \"23fb3f29-8c86-4ba7-8260-4ead78279db6\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.570079 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56z99\" (UniqueName: \"kubernetes.io/projected/23fb3f29-8c86-4ba7-8260-4ead78279db6-kube-api-access-56z99\") pod \"23fb3f29-8c86-4ba7-8260-4ead78279db6\" (UID: \"23fb3f29-8c86-4ba7-8260-4ead78279db6\") " Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.577419 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23fb3f29-8c86-4ba7-8260-4ead78279db6-kube-api-access-56z99" (OuterVolumeSpecName: "kube-api-access-56z99") pod "23fb3f29-8c86-4ba7-8260-4ead78279db6" (UID: "23fb3f29-8c86-4ba7-8260-4ead78279db6"). InnerVolumeSpecName "kube-api-access-56z99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.599224 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-config-data" (OuterVolumeSpecName: "config-data") pod "23fb3f29-8c86-4ba7-8260-4ead78279db6" (UID: "23fb3f29-8c86-4ba7-8260-4ead78279db6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.603459 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23fb3f29-8c86-4ba7-8260-4ead78279db6" (UID: "23fb3f29-8c86-4ba7-8260-4ead78279db6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.642982 4696 generic.go:334] "Generic (PLEG): container finished" podID="23fb3f29-8c86-4ba7-8260-4ead78279db6" containerID="3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b" exitCode=137 Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.643023 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"23fb3f29-8c86-4ba7-8260-4ead78279db6","Type":"ContainerDied","Data":"3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b"} Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.643065 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"23fb3f29-8c86-4ba7-8260-4ead78279db6","Type":"ContainerDied","Data":"34674ceea1ab5f008225bda043843b18c0d4253a513a34b884821e553e04e7cc"} Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.643085 4696 scope.go:117] "RemoveContainer" containerID="3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.643081 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.672710 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.673201 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23fb3f29-8c86-4ba7-8260-4ead78279db6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.673223 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56z99\" (UniqueName: \"kubernetes.io/projected/23fb3f29-8c86-4ba7-8260-4ead78279db6-kube-api-access-56z99\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.706746 4696 scope.go:117] "RemoveContainer" containerID="3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b" Nov 25 10:58:03 crc kubenswrapper[4696]: E1125 10:58:03.709143 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b\": container with ID starting with 3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b not found: ID does not exist" containerID="3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.709186 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b"} err="failed to get container status \"3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b\": rpc error: code = NotFound desc = could not find container \"3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b\": container with ID starting with 3669f57422a0ff90f93529fc4bb2c8baafbc84f007bf4499bf11e07623ef626b not found: ID does not exist" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.711915 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.726262 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.766950 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:58:03 crc kubenswrapper[4696]: E1125 10:58:03.767452 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fb3f29-8c86-4ba7-8260-4ead78279db6" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.767473 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fb3f29-8c86-4ba7-8260-4ead78279db6" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.767739 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="23fb3f29-8c86-4ba7-8260-4ead78279db6" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.768439 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.778883 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.779303 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.779368 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.779905 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.779701 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.780103 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.780169 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6w78\" (UniqueName: \"kubernetes.io/projected/672502a7-5ef3-4596-9507-21afa27d44e6-kube-api-access-t6w78\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.780206 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.786448 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.881099 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.881163 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.881212 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.881238 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6w78\" (UniqueName: \"kubernetes.io/projected/672502a7-5ef3-4596-9507-21afa27d44e6-kube-api-access-t6w78\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.881256 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.885491 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.885830 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.885997 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.886053 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672502a7-5ef3-4596-9507-21afa27d44e6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:03 crc kubenswrapper[4696]: I1125 10:58:03.897635 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6w78\" (UniqueName: \"kubernetes.io/projected/672502a7-5ef3-4596-9507-21afa27d44e6-kube-api-access-t6w78\") pod \"nova-cell1-novncproxy-0\" (UID: \"672502a7-5ef3-4596-9507-21afa27d44e6\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:04 crc kubenswrapper[4696]: I1125 10:58:04.059449 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23fb3f29-8c86-4ba7-8260-4ead78279db6" path="/var/lib/kubelet/pods/23fb3f29-8c86-4ba7-8260-4ead78279db6/volumes" Nov 25 10:58:04 crc kubenswrapper[4696]: I1125 10:58:04.093344 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:04 crc kubenswrapper[4696]: I1125 10:58:04.582798 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:58:04 crc kubenswrapper[4696]: W1125 10:58:04.599926 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod672502a7_5ef3_4596_9507_21afa27d44e6.slice/crio-d93067c36f4a08612bf9f1d5c92c46dca297a40c995c3559d9f6b76ee7b01cfa WatchSource:0}: Error finding container d93067c36f4a08612bf9f1d5c92c46dca297a40c995c3559d9f6b76ee7b01cfa: Status 404 returned error can't find the container with id d93067c36f4a08612bf9f1d5c92c46dca297a40c995c3559d9f6b76ee7b01cfa Nov 25 10:58:04 crc kubenswrapper[4696]: I1125 10:58:04.655779 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"672502a7-5ef3-4596-9507-21afa27d44e6","Type":"ContainerStarted","Data":"d93067c36f4a08612bf9f1d5c92c46dca297a40c995c3559d9f6b76ee7b01cfa"} Nov 25 10:58:05 crc kubenswrapper[4696]: I1125 10:58:05.667919 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"672502a7-5ef3-4596-9507-21afa27d44e6","Type":"ContainerStarted","Data":"31596c5b3f16e73556ac15f8642c8b3a2a1278f897388204cebfef89906f350c"} Nov 25 10:58:05 crc kubenswrapper[4696]: I1125 10:58:05.689762 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.689741181 podStartE2EDuration="2.689741181s" podCreationTimestamp="2025-11-25 10:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:05.687313945 +0000 UTC m=+1282.500930562" watchObservedRunningTime="2025-11-25 10:58:05.689741181 +0000 UTC m=+1282.503357768" Nov 25 10:58:06 crc kubenswrapper[4696]: I1125 10:58:06.862016 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:58:06 crc kubenswrapper[4696]: I1125 10:58:06.862853 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:58:06 crc kubenswrapper[4696]: I1125 10:58:06.864704 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:58:06 crc kubenswrapper[4696]: I1125 10:58:06.866527 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.687204 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.690913 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.898450 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-kbdbq"] Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.904885 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.933981 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-kbdbq"] Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.953263 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znhr6\" (UniqueName: \"kubernetes.io/projected/bb61a07b-9c8f-4f67-b714-05d82b3b1715-kube-api-access-znhr6\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.953344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.953492 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.953532 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-config\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.953559 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:07 crc kubenswrapper[4696]: I1125 10:58:07.953749 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.055929 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.056019 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znhr6\" (UniqueName: \"kubernetes.io/projected/bb61a07b-9c8f-4f67-b714-05d82b3b1715-kube-api-access-znhr6\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.056052 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.056114 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.056139 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-config\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.056165 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.057010 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.080653 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.080693 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.081196 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.084716 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-config\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.108786 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znhr6\" (UniqueName: \"kubernetes.io/projected/bb61a07b-9c8f-4f67-b714-05d82b3b1715-kube-api-access-znhr6\") pod \"dnsmasq-dns-cd5cbd7b9-kbdbq\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.229460 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:08 crc kubenswrapper[4696]: W1125 10:58:08.732247 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb61a07b_9c8f_4f67_b714_05d82b3b1715.slice/crio-cb382a9190fd086b53aea6980331b723bed7ee56ec8da81dcb10cddf2b6637cf WatchSource:0}: Error finding container cb382a9190fd086b53aea6980331b723bed7ee56ec8da81dcb10cddf2b6637cf: Status 404 returned error can't find the container with id cb382a9190fd086b53aea6980331b723bed7ee56ec8da81dcb10cddf2b6637cf Nov 25 10:58:08 crc kubenswrapper[4696]: I1125 10:58:08.735657 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-kbdbq"] Nov 25 10:58:09 crc kubenswrapper[4696]: I1125 10:58:09.095155 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:09 crc kubenswrapper[4696]: I1125 10:58:09.705326 4696 generic.go:334] "Generic (PLEG): container finished" podID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerID="2882758aa48311ddca88e2b788cea603b9658e5c7eef2bd0410a41d74f97dd4b" exitCode=0 Nov 25 10:58:09 crc kubenswrapper[4696]: I1125 10:58:09.705383 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" event={"ID":"bb61a07b-9c8f-4f67-b714-05d82b3b1715","Type":"ContainerDied","Data":"2882758aa48311ddca88e2b788cea603b9658e5c7eef2bd0410a41d74f97dd4b"} Nov 25 10:58:09 crc kubenswrapper[4696]: I1125 10:58:09.705679 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" event={"ID":"bb61a07b-9c8f-4f67-b714-05d82b3b1715","Type":"ContainerStarted","Data":"cb382a9190fd086b53aea6980331b723bed7ee56ec8da81dcb10cddf2b6637cf"} Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.309898 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.310429 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-central-agent" containerID="cri-o://0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb" gracePeriod=30 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.311221 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-notification-agent" containerID="cri-o://f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177" gracePeriod=30 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.311233 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="sg-core" containerID="cri-o://659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb" gracePeriod=30 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.311404 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="proxy-httpd" containerID="cri-o://f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330" gracePeriod=30 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.417162 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": read tcp 10.217.0.2:58320->10.217.0.196:3000: read: connection reset by peer" Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.485512 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.716097 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" event={"ID":"bb61a07b-9c8f-4f67-b714-05d82b3b1715","Type":"ContainerStarted","Data":"8525bc7bbac5dc4f6924ad859e4b8311e5a96782b323e57bd3f56f9e34cf1404"} Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.716351 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.720176 4696 generic.go:334] "Generic (PLEG): container finished" podID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerID="f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330" exitCode=0 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.720208 4696 generic.go:334] "Generic (PLEG): container finished" podID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerID="659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb" exitCode=2 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.720388 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-log" containerID="cri-o://2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a" gracePeriod=30 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.720603 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerDied","Data":"f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330"} Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.720628 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerDied","Data":"659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb"} Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.720679 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-api" containerID="cri-o://71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc" gracePeriod=30 Nov 25 10:58:10 crc kubenswrapper[4696]: I1125 10:58:10.744417 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" podStartSLOduration=3.744395217 podStartE2EDuration="3.744395217s" podCreationTimestamp="2025-11-25 10:58:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:10.735341091 +0000 UTC m=+1287.548957678" watchObservedRunningTime="2025-11-25 10:58:10.744395217 +0000 UTC m=+1287.558011824" Nov 25 10:58:11 crc kubenswrapper[4696]: I1125 10:58:11.731174 4696 generic.go:334] "Generic (PLEG): container finished" podID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerID="0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb" exitCode=0 Nov 25 10:58:11 crc kubenswrapper[4696]: I1125 10:58:11.731413 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerDied","Data":"0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb"} Nov 25 10:58:11 crc kubenswrapper[4696]: I1125 10:58:11.733492 4696 generic.go:334] "Generic (PLEG): container finished" podID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerID="2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a" exitCode=143 Nov 25 10:58:11 crc kubenswrapper[4696]: I1125 10:58:11.733562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3041af3a-3764-4e9d-aedf-c45ef8e34edf","Type":"ContainerDied","Data":"2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a"} Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.094564 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.112702 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.406960 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.413404 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491253 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-log-httpd\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491325 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql8mz\" (UniqueName: \"kubernetes.io/projected/3041af3a-3764-4e9d-aedf-c45ef8e34edf-kube-api-access-ql8mz\") pod \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491346 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041af3a-3764-4e9d-aedf-c45ef8e34edf-logs\") pod \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491384 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zm6r\" (UniqueName: \"kubernetes.io/projected/98853f51-ccc5-40df-8f01-00f2fabe8668-kube-api-access-5zm6r\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491484 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-config-data\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491522 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-config-data\") pod \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491551 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-ceilometer-tls-certs\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491571 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-scripts\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491600 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-sg-core-conf-yaml\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491621 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-run-httpd\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491675 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-combined-ca-bundle\") pod \"98853f51-ccc5-40df-8f01-00f2fabe8668\" (UID: \"98853f51-ccc5-40df-8f01-00f2fabe8668\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.491784 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-combined-ca-bundle\") pod \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\" (UID: \"3041af3a-3764-4e9d-aedf-c45ef8e34edf\") " Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.496560 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.500050 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3041af3a-3764-4e9d-aedf-c45ef8e34edf-logs" (OuterVolumeSpecName: "logs") pod "3041af3a-3764-4e9d-aedf-c45ef8e34edf" (UID: "3041af3a-3764-4e9d-aedf-c45ef8e34edf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.500716 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.507400 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-scripts" (OuterVolumeSpecName: "scripts") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.513738 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98853f51-ccc5-40df-8f01-00f2fabe8668-kube-api-access-5zm6r" (OuterVolumeSpecName: "kube-api-access-5zm6r") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "kube-api-access-5zm6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.532472 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3041af3a-3764-4e9d-aedf-c45ef8e34edf-kube-api-access-ql8mz" (OuterVolumeSpecName: "kube-api-access-ql8mz") pod "3041af3a-3764-4e9d-aedf-c45ef8e34edf" (UID: "3041af3a-3764-4e9d-aedf-c45ef8e34edf"). InnerVolumeSpecName "kube-api-access-ql8mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.594037 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.594074 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.594097 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98853f51-ccc5-40df-8f01-00f2fabe8668-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.594108 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql8mz\" (UniqueName: \"kubernetes.io/projected/3041af3a-3764-4e9d-aedf-c45ef8e34edf-kube-api-access-ql8mz\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.594121 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041af3a-3764-4e9d-aedf-c45ef8e34edf-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.594131 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zm6r\" (UniqueName: \"kubernetes.io/projected/98853f51-ccc5-40df-8f01-00f2fabe8668-kube-api-access-5zm6r\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.622440 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-config-data" (OuterVolumeSpecName: "config-data") pod "3041af3a-3764-4e9d-aedf-c45ef8e34edf" (UID: "3041af3a-3764-4e9d-aedf-c45ef8e34edf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.623182 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.639900 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3041af3a-3764-4e9d-aedf-c45ef8e34edf" (UID: "3041af3a-3764-4e9d-aedf-c45ef8e34edf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.651814 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.696239 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.696668 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.698886 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.698908 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041af3a-3764-4e9d-aedf-c45ef8e34edf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.717878 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.725398 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-config-data" (OuterVolumeSpecName: "config-data") pod "98853f51-ccc5-40df-8f01-00f2fabe8668" (UID: "98853f51-ccc5-40df-8f01-00f2fabe8668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.764523 4696 generic.go:334] "Generic (PLEG): container finished" podID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerID="f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177" exitCode=0 Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.764586 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerDied","Data":"f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177"} Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.764612 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98853f51-ccc5-40df-8f01-00f2fabe8668","Type":"ContainerDied","Data":"1831648904a1d1122bf85ce2c4321836792ec5239d1647b95f750be7cee2c92b"} Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.764629 4696 scope.go:117] "RemoveContainer" containerID="f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.764875 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.774266 4696 generic.go:334] "Generic (PLEG): container finished" podID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerID="71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc" exitCode=0 Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.774423 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.774461 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3041af3a-3764-4e9d-aedf-c45ef8e34edf","Type":"ContainerDied","Data":"71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc"} Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.774546 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3041af3a-3764-4e9d-aedf-c45ef8e34edf","Type":"ContainerDied","Data":"f0932dcb782465230d1609f5ea390a9e8e0086ff5b4e8cbe210974575faf723c"} Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.801059 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.801093 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98853f51-ccc5-40df-8f01-00f2fabe8668-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.809685 4696 scope.go:117] "RemoveContainer" containerID="659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.819410 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.827545 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.839049 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.847762 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.848637 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="sg-core" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.848702 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="sg-core" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.848730 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-notification-agent" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.848738 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-notification-agent" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.848787 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-api" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.848797 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-api" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.848818 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="proxy-httpd" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.848825 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="proxy-httpd" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.848866 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-log" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.848876 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-log" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.848902 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-central-agent" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.848910 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-central-agent" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.849479 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="sg-core" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.853497 4696 scope.go:117] "RemoveContainer" containerID="f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.855069 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-log" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.855126 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="proxy-httpd" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.855139 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" containerName="nova-api-api" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.855168 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-notification-agent" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.855750 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" containerName="ceilometer-central-agent" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.868958 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.869087 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.882479 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.884653 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.884865 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.885003 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.901754 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.902518 4696 scope.go:117] "RemoveContainer" containerID="0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.929497 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.930915 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.934581 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.934588 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.934782 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.941185 4696 scope.go:117] "RemoveContainer" containerID="f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.942236 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330\": container with ID starting with f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330 not found: ID does not exist" containerID="f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.942550 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330"} err="failed to get container status \"f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330\": rpc error: code = NotFound desc = could not find container \"f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330\": container with ID starting with f92e7d0c64fb3ac302e338e7f9ac62c48f2e2fcf5a56c8b851ea6855b59ed330 not found: ID does not exist" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.942578 4696 scope.go:117] "RemoveContainer" containerID="659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.943322 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb\": container with ID starting with 659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb not found: ID does not exist" containerID="659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.943646 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb"} err="failed to get container status \"659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb\": rpc error: code = NotFound desc = could not find container \"659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb\": container with ID starting with 659813e2e4bacb66134782333d30187e37db7fafa96f5a430490bb6da9599deb not found: ID does not exist" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.943727 4696 scope.go:117] "RemoveContainer" containerID="f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.944153 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177\": container with ID starting with f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177 not found: ID does not exist" containerID="f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.944495 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177"} err="failed to get container status \"f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177\": rpc error: code = NotFound desc = could not find container \"f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177\": container with ID starting with f9ad54bd144ae95d7b2be42e48bda94be9c4916364e6ee7d9cea15b40bcdb177 not found: ID does not exist" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.944517 4696 scope.go:117] "RemoveContainer" containerID="0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb" Nov 25 10:58:14 crc kubenswrapper[4696]: E1125 10:58:14.945103 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb\": container with ID starting with 0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb not found: ID does not exist" containerID="0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.945149 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb"} err="failed to get container status \"0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb\": rpc error: code = NotFound desc = could not find container \"0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb\": container with ID starting with 0157ca1cf00ce28c8f3d3c0fe17ac43ba0c1194579d234cfce6f881d73d8a6fb not found: ID does not exist" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.945325 4696 scope.go:117] "RemoveContainer" containerID="71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc" Nov 25 10:58:14 crc kubenswrapper[4696]: I1125 10:58:14.965820 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.005792 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.005873 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-scripts\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.005902 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.005954 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-run-httpd\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.005975 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-config-data\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.006014 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.006057 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvbnz\" (UniqueName: \"kubernetes.io/projected/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-kube-api-access-wvbnz\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.006079 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-log-httpd\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.039034 4696 scope.go:117] "RemoveContainer" containerID="2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.079471 4696 scope.go:117] "RemoveContainer" containerID="71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc" Nov 25 10:58:15 crc kubenswrapper[4696]: E1125 10:58:15.079931 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc\": container with ID starting with 71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc not found: ID does not exist" containerID="71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.079956 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc"} err="failed to get container status \"71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc\": rpc error: code = NotFound desc = could not find container \"71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc\": container with ID starting with 71073500458dca353977aceef47a868565ecbdb0f52b52fac9051e6afca927bc not found: ID does not exist" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.079975 4696 scope.go:117] "RemoveContainer" containerID="2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a" Nov 25 10:58:15 crc kubenswrapper[4696]: E1125 10:58:15.082777 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a\": container with ID starting with 2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a not found: ID does not exist" containerID="2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.082810 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a"} err="failed to get container status \"2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a\": rpc error: code = NotFound desc = could not find container \"2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a\": container with ID starting with 2d702a38b5da48fad2ac48757bf56159b1640767aaebe68d043b6555f48d345a not found: ID does not exist" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.107822 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108114 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-public-tls-certs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108137 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-config-data\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108182 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvbnz\" (UniqueName: \"kubernetes.io/projected/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-kube-api-access-wvbnz\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108215 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-log-httpd\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108242 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108283 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108317 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-scripts\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108340 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108388 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-logs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108402 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108422 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-run-httpd\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108438 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lssnq\" (UniqueName: \"kubernetes.io/projected/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-kube-api-access-lssnq\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.108455 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-config-data\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.110244 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-log-httpd\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.116013 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-run-httpd\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.121634 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.123835 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.124930 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-scripts\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.128301 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-config-data\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.141690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.143173 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvbnz\" (UniqueName: \"kubernetes.io/projected/9fa6192b-7ea2-414b-9e4f-ebb5648a58bf-kube-api-access-wvbnz\") pod \"ceilometer-0\" (UID: \"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf\") " pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.145535 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2s8v9"] Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.147086 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.152598 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.152917 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.163850 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2s8v9"] Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.206919 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.209913 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-public-tls-certs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.210443 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-config-data\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.210801 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.210965 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-logs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.211040 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.211122 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lssnq\" (UniqueName: \"kubernetes.io/projected/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-kube-api-access-lssnq\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.212045 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-logs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.213629 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-public-tls-certs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.217311 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-config-data\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.218917 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.219155 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.228195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lssnq\" (UniqueName: \"kubernetes.io/projected/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-kube-api-access-lssnq\") pod \"nova-api-0\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.312260 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.312303 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-scripts\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.312322 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-config-data\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.312459 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn6fw\" (UniqueName: \"kubernetes.io/projected/06860896-de50-41b9-9770-156967ac658f-kube-api-access-hn6fw\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.330099 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.414424 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.414767 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-scripts\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.414794 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-config-data\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.414985 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn6fw\" (UniqueName: \"kubernetes.io/projected/06860896-de50-41b9-9770-156967ac658f-kube-api-access-hn6fw\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.428195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-config-data\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.428376 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-scripts\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.428487 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.436736 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn6fw\" (UniqueName: \"kubernetes.io/projected/06860896-de50-41b9-9770-156967ac658f-kube-api-access-hn6fw\") pod \"nova-cell1-cell-mapping-2s8v9\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.635134 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.712209 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.787318 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf","Type":"ContainerStarted","Data":"ffab6d4780e41453e0ff2e210f8a25aa3bb950638444c5ee76859fda1c8e7b87"} Nov 25 10:58:15 crc kubenswrapper[4696]: I1125 10:58:15.847058 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.057745 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3041af3a-3764-4e9d-aedf-c45ef8e34edf" path="/var/lib/kubelet/pods/3041af3a-3764-4e9d-aedf-c45ef8e34edf/volumes" Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.058434 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98853f51-ccc5-40df-8f01-00f2fabe8668" path="/var/lib/kubelet/pods/98853f51-ccc5-40df-8f01-00f2fabe8668/volumes" Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.098369 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2s8v9"] Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.797822 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf","Type":"ContainerStarted","Data":"e62e291f3a701633ae7fea81a7a5b2ac986ac071335fcf3ecf5877aca36cc964"} Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.800740 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2s8v9" event={"ID":"06860896-de50-41b9-9770-156967ac658f","Type":"ContainerStarted","Data":"763a7b7a7e164083e8a32a3a178cea38e62790734ded56d23a26b1b113345c41"} Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.800804 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2s8v9" event={"ID":"06860896-de50-41b9-9770-156967ac658f","Type":"ContainerStarted","Data":"2c3458d8db1e14d9fe568347f6f9dc7723afc498ac3dc4616dce448afc678595"} Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.804636 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1","Type":"ContainerStarted","Data":"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63"} Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.804690 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1","Type":"ContainerStarted","Data":"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303"} Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.804703 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1","Type":"ContainerStarted","Data":"c5fff2ce875ee3c0cff5513224181a32abdad15600f1a54abe10bc3597e1ef50"} Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.819323 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2s8v9" podStartSLOduration=1.81930311 podStartE2EDuration="1.81930311s" podCreationTimestamp="2025-11-25 10:58:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:16.816497252 +0000 UTC m=+1293.630113839" watchObservedRunningTime="2025-11-25 10:58:16.81930311 +0000 UTC m=+1293.632919697" Nov 25 10:58:16 crc kubenswrapper[4696]: I1125 10:58:16.838445 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8384255940000003 podStartE2EDuration="2.838425594s" podCreationTimestamp="2025-11-25 10:58:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:16.835581904 +0000 UTC m=+1293.649198501" watchObservedRunningTime="2025-11-25 10:58:16.838425594 +0000 UTC m=+1293.652042181" Nov 25 10:58:17 crc kubenswrapper[4696]: I1125 10:58:17.824844 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf","Type":"ContainerStarted","Data":"26512090aef2fd0a17e49b676fc75c54babe0315b5c009ff0fde0bdfa4bcb56c"} Nov 25 10:58:17 crc kubenswrapper[4696]: I1125 10:58:17.825419 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf","Type":"ContainerStarted","Data":"961e9458bcec1192384f01cc10fb201e4019c89ad6092b548d5c48e17a8b089b"} Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.230797 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.314041 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4wvx4"] Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.314586 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerName="dnsmasq-dns" containerID="cri-o://30810c9e3eecbf9e1be83ee7ada1ac9b5ccadf967e125fb72e9068e4d53c64f7" gracePeriod=10 Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.840006 4696 generic.go:334] "Generic (PLEG): container finished" podID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerID="30810c9e3eecbf9e1be83ee7ada1ac9b5ccadf967e125fb72e9068e4d53c64f7" exitCode=0 Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.840049 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" event={"ID":"15f521d2-8aef-4ea6-b26d-1e2a259477ee","Type":"ContainerDied","Data":"30810c9e3eecbf9e1be83ee7ada1ac9b5ccadf967e125fb72e9068e4d53c64f7"} Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.840074 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" event={"ID":"15f521d2-8aef-4ea6-b26d-1e2a259477ee","Type":"ContainerDied","Data":"90ec1f9ca882c89ea4e00222a652d851f458daa620b7d64bc02c9c41d62b5406"} Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.840091 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90ec1f9ca882c89ea4e00222a652d851f458daa620b7d64bc02c9c41d62b5406" Nov 25 10:58:18 crc kubenswrapper[4696]: I1125 10:58:18.917324 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.086920 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-config\") pod \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.086995 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-sb\") pod \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.087045 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-svc\") pod \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.087618 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-swift-storage-0\") pod \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.087728 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfwft\" (UniqueName: \"kubernetes.io/projected/15f521d2-8aef-4ea6-b26d-1e2a259477ee-kube-api-access-qfwft\") pod \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.087757 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-nb\") pod \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\" (UID: \"15f521d2-8aef-4ea6-b26d-1e2a259477ee\") " Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.112074 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f521d2-8aef-4ea6-b26d-1e2a259477ee-kube-api-access-qfwft" (OuterVolumeSpecName: "kube-api-access-qfwft") pod "15f521d2-8aef-4ea6-b26d-1e2a259477ee" (UID: "15f521d2-8aef-4ea6-b26d-1e2a259477ee"). InnerVolumeSpecName "kube-api-access-qfwft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.152533 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "15f521d2-8aef-4ea6-b26d-1e2a259477ee" (UID: "15f521d2-8aef-4ea6-b26d-1e2a259477ee"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.177770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15f521d2-8aef-4ea6-b26d-1e2a259477ee" (UID: "15f521d2-8aef-4ea6-b26d-1e2a259477ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.179631 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15f521d2-8aef-4ea6-b26d-1e2a259477ee" (UID: "15f521d2-8aef-4ea6-b26d-1e2a259477ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.182386 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-config" (OuterVolumeSpecName: "config") pod "15f521d2-8aef-4ea6-b26d-1e2a259477ee" (UID: "15f521d2-8aef-4ea6-b26d-1e2a259477ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.190181 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.190199 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.190209 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.190219 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfwft\" (UniqueName: \"kubernetes.io/projected/15f521d2-8aef-4ea6-b26d-1e2a259477ee-kube-api-access-qfwft\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.190229 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.193360 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15f521d2-8aef-4ea6-b26d-1e2a259477ee" (UID: "15f521d2-8aef-4ea6-b26d-1e2a259477ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.291974 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15f521d2-8aef-4ea6-b26d-1e2a259477ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.850718 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4wvx4" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.857877 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fa6192b-7ea2-414b-9e4f-ebb5648a58bf","Type":"ContainerStarted","Data":"092081b635f17a23b92762d9f870d1d017a6367d85b92b9d8cb0643e2cfb4508"} Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.858442 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.888808 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.894186842 podStartE2EDuration="5.888785212s" podCreationTimestamp="2025-11-25 10:58:14 +0000 UTC" firstStartedPulling="2025-11-25 10:58:15.731538231 +0000 UTC m=+1292.545154818" lastFinishedPulling="2025-11-25 10:58:18.726136601 +0000 UTC m=+1295.539753188" observedRunningTime="2025-11-25 10:58:19.877027322 +0000 UTC m=+1296.690643929" watchObservedRunningTime="2025-11-25 10:58:19.888785212 +0000 UTC m=+1296.702401799" Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.908275 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4wvx4"] Nov 25 10:58:19 crc kubenswrapper[4696]: I1125 10:58:19.917159 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4wvx4"] Nov 25 10:58:20 crc kubenswrapper[4696]: I1125 10:58:20.057763 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" path="/var/lib/kubelet/pods/15f521d2-8aef-4ea6-b26d-1e2a259477ee/volumes" Nov 25 10:58:21 crc kubenswrapper[4696]: I1125 10:58:21.870653 4696 generic.go:334] "Generic (PLEG): container finished" podID="06860896-de50-41b9-9770-156967ac658f" containerID="763a7b7a7e164083e8a32a3a178cea38e62790734ded56d23a26b1b113345c41" exitCode=0 Nov 25 10:58:21 crc kubenswrapper[4696]: I1125 10:58:21.870725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2s8v9" event={"ID":"06860896-de50-41b9-9770-156967ac658f","Type":"ContainerDied","Data":"763a7b7a7e164083e8a32a3a178cea38e62790734ded56d23a26b1b113345c41"} Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.277896 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.381580 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-config-data\") pod \"06860896-de50-41b9-9770-156967ac658f\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.381654 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-combined-ca-bundle\") pod \"06860896-de50-41b9-9770-156967ac658f\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.381913 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn6fw\" (UniqueName: \"kubernetes.io/projected/06860896-de50-41b9-9770-156967ac658f-kube-api-access-hn6fw\") pod \"06860896-de50-41b9-9770-156967ac658f\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.381972 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-scripts\") pod \"06860896-de50-41b9-9770-156967ac658f\" (UID: \"06860896-de50-41b9-9770-156967ac658f\") " Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.388170 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-scripts" (OuterVolumeSpecName: "scripts") pod "06860896-de50-41b9-9770-156967ac658f" (UID: "06860896-de50-41b9-9770-156967ac658f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.389246 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06860896-de50-41b9-9770-156967ac658f-kube-api-access-hn6fw" (OuterVolumeSpecName: "kube-api-access-hn6fw") pod "06860896-de50-41b9-9770-156967ac658f" (UID: "06860896-de50-41b9-9770-156967ac658f"). InnerVolumeSpecName "kube-api-access-hn6fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.414749 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06860896-de50-41b9-9770-156967ac658f" (UID: "06860896-de50-41b9-9770-156967ac658f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.415106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-config-data" (OuterVolumeSpecName: "config-data") pod "06860896-de50-41b9-9770-156967ac658f" (UID: "06860896-de50-41b9-9770-156967ac658f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.484077 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn6fw\" (UniqueName: \"kubernetes.io/projected/06860896-de50-41b9-9770-156967ac658f-kube-api-access-hn6fw\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.484126 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.484138 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.484146 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06860896-de50-41b9-9770-156967ac658f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.892855 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2s8v9" event={"ID":"06860896-de50-41b9-9770-156967ac658f","Type":"ContainerDied","Data":"2c3458d8db1e14d9fe568347f6f9dc7723afc498ac3dc4616dce448afc678595"} Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.893710 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c3458d8db1e14d9fe568347f6f9dc7723afc498ac3dc4616dce448afc678595" Nov 25 10:58:23 crc kubenswrapper[4696]: I1125 10:58:23.893869 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2s8v9" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.070563 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.070838 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-log" containerID="cri-o://c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303" gracePeriod=30 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.071145 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-api" containerID="cri-o://5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63" gracePeriod=30 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.089341 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.089567 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c6f46bac-b868-4767-8efb-5fdd89267c66" containerName="nova-scheduler-scheduler" containerID="cri-o://e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1" gracePeriod=30 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.100435 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.100729 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-log" containerID="cri-o://318d4f95c617c7edcb581ef02df1cde2006040322c6877f5f5babd4e33a4ef63" gracePeriod=30 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.100862 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-metadata" containerID="cri-o://a1e21147c6240a548d496dccb577f9e853613f365e7d7265b9d79e18445e106e" gracePeriod=30 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.842579 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932479 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1","Type":"ContainerDied","Data":"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63"} Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932552 4696 scope.go:117] "RemoveContainer" containerID="5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932516 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932752 4696 generic.go:334] "Generic (PLEG): container finished" podID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerID="5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63" exitCode=0 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932772 4696 generic.go:334] "Generic (PLEG): container finished" podID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerID="c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303" exitCode=143 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932820 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1","Type":"ContainerDied","Data":"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303"} Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.932832 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1","Type":"ContainerDied","Data":"c5fff2ce875ee3c0cff5513224181a32abdad15600f1a54abe10bc3597e1ef50"} Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.945006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b67cc31d-f86a-4851-9ac2-4936d79f8273","Type":"ContainerDied","Data":"318d4f95c617c7edcb581ef02df1cde2006040322c6877f5f5babd4e33a4ef63"} Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.945057 4696 generic.go:334] "Generic (PLEG): container finished" podID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerID="318d4f95c617c7edcb581ef02df1cde2006040322c6877f5f5babd4e33a4ef63" exitCode=143 Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.966682 4696 scope.go:117] "RemoveContainer" containerID="c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.990287 4696 scope.go:117] "RemoveContainer" containerID="5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63" Nov 25 10:58:24 crc kubenswrapper[4696]: E1125 10:58:24.991154 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63\": container with ID starting with 5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63 not found: ID does not exist" containerID="5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.991205 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63"} err="failed to get container status \"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63\": rpc error: code = NotFound desc = could not find container \"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63\": container with ID starting with 5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63 not found: ID does not exist" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.991228 4696 scope.go:117] "RemoveContainer" containerID="c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303" Nov 25 10:58:24 crc kubenswrapper[4696]: E1125 10:58:24.991636 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303\": container with ID starting with c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303 not found: ID does not exist" containerID="c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.991698 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303"} err="failed to get container status \"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303\": rpc error: code = NotFound desc = could not find container \"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303\": container with ID starting with c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303 not found: ID does not exist" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.991720 4696 scope.go:117] "RemoveContainer" containerID="5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.992050 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63"} err="failed to get container status \"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63\": rpc error: code = NotFound desc = could not find container \"5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63\": container with ID starting with 5331a47893df35ec0ba168c33f7e76ae9aedb4346dddee1bf287e57b6452ae63 not found: ID does not exist" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.992085 4696 scope.go:117] "RemoveContainer" containerID="c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303" Nov 25 10:58:24 crc kubenswrapper[4696]: I1125 10:58:24.992341 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303"} err="failed to get container status \"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303\": rpc error: code = NotFound desc = could not find container \"c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303\": container with ID starting with c611887f8cbea8aa32305856509c43b4d14d01d38da73d226be6c26aca8d7303 not found: ID does not exist" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012255 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lssnq\" (UniqueName: \"kubernetes.io/projected/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-kube-api-access-lssnq\") pod \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012306 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-public-tls-certs\") pod \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012338 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-internal-tls-certs\") pod \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012430 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-logs\") pod \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012516 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-config-data\") pod \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012573 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-combined-ca-bundle\") pod \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\" (UID: \"dfa5f437-b259-4cd6-8044-e3c1aa7f56a1\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.012782 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-logs" (OuterVolumeSpecName: "logs") pod "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" (UID: "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.013131 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.018351 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-kube-api-access-lssnq" (OuterVolumeSpecName: "kube-api-access-lssnq") pod "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" (UID: "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1"). InnerVolumeSpecName "kube-api-access-lssnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.047979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" (UID: "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.050831 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-config-data" (OuterVolumeSpecName: "config-data") pod "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" (UID: "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.077172 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" (UID: "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.078296 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" (UID: "dfa5f437-b259-4cd6-8044-e3c1aa7f56a1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.116296 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.116330 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.116343 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lssnq\" (UniqueName: \"kubernetes.io/projected/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-kube-api-access-lssnq\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.116352 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.116362 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.271547 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.291513 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313134 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:25 crc kubenswrapper[4696]: E1125 10:58:25.313515 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06860896-de50-41b9-9770-156967ac658f" containerName="nova-manage" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313526 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="06860896-de50-41b9-9770-156967ac658f" containerName="nova-manage" Nov 25 10:58:25 crc kubenswrapper[4696]: E1125 10:58:25.313542 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-log" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313548 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-log" Nov 25 10:58:25 crc kubenswrapper[4696]: E1125 10:58:25.313574 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerName="dnsmasq-dns" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313580 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerName="dnsmasq-dns" Nov 25 10:58:25 crc kubenswrapper[4696]: E1125 10:58:25.313588 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-api" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313594 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-api" Nov 25 10:58:25 crc kubenswrapper[4696]: E1125 10:58:25.313605 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerName="init" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313610 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerName="init" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313833 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f521d2-8aef-4ea6-b26d-1e2a259477ee" containerName="dnsmasq-dns" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313846 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="06860896-de50-41b9-9770-156967ac658f" containerName="nova-manage" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313862 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-api" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.313878 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" containerName="nova-api-log" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.314812 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.323690 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.323898 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.324008 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.331899 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.425294 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-public-tls-certs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.425722 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l89hc\" (UniqueName: \"kubernetes.io/projected/8377dca6-a47c-4ffb-9916-f6505eed1e48-kube-api-access-l89hc\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.425816 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8377dca6-a47c-4ffb-9916-f6505eed1e48-logs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.425885 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-config-data\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.425915 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.425939 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.527895 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8377dca6-a47c-4ffb-9916-f6505eed1e48-logs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.527953 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-config-data\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.527976 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.527994 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.528045 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-public-tls-certs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.528097 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l89hc\" (UniqueName: \"kubernetes.io/projected/8377dca6-a47c-4ffb-9916-f6505eed1e48-kube-api-access-l89hc\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.528327 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8377dca6-a47c-4ffb-9916-f6505eed1e48-logs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.543301 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.543309 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-public-tls-certs\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.543401 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-config-data\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.554611 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8377dca6-a47c-4ffb-9916-f6505eed1e48-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.562203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l89hc\" (UniqueName: \"kubernetes.io/projected/8377dca6-a47c-4ffb-9916-f6505eed1e48-kube-api-access-l89hc\") pod \"nova-api-0\" (UID: \"8377dca6-a47c-4ffb-9916-f6505eed1e48\") " pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.678055 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.742841 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.834179 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lchpx\" (UniqueName: \"kubernetes.io/projected/c6f46bac-b868-4767-8efb-5fdd89267c66-kube-api-access-lchpx\") pod \"c6f46bac-b868-4767-8efb-5fdd89267c66\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.834635 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-combined-ca-bundle\") pod \"c6f46bac-b868-4767-8efb-5fdd89267c66\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.834680 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-config-data\") pod \"c6f46bac-b868-4767-8efb-5fdd89267c66\" (UID: \"c6f46bac-b868-4767-8efb-5fdd89267c66\") " Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.845174 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6f46bac-b868-4767-8efb-5fdd89267c66-kube-api-access-lchpx" (OuterVolumeSpecName: "kube-api-access-lchpx") pod "c6f46bac-b868-4767-8efb-5fdd89267c66" (UID: "c6f46bac-b868-4767-8efb-5fdd89267c66"). InnerVolumeSpecName "kube-api-access-lchpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.867944 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-config-data" (OuterVolumeSpecName: "config-data") pod "c6f46bac-b868-4767-8efb-5fdd89267c66" (UID: "c6f46bac-b868-4767-8efb-5fdd89267c66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.868037 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6f46bac-b868-4767-8efb-5fdd89267c66" (UID: "c6f46bac-b868-4767-8efb-5fdd89267c66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.937314 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lchpx\" (UniqueName: \"kubernetes.io/projected/c6f46bac-b868-4767-8efb-5fdd89267c66-kube-api-access-lchpx\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.937349 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.937360 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f46bac-b868-4767-8efb-5fdd89267c66-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.961564 4696 generic.go:334] "Generic (PLEG): container finished" podID="c6f46bac-b868-4767-8efb-5fdd89267c66" containerID="e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1" exitCode=0 Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.961603 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6f46bac-b868-4767-8efb-5fdd89267c66","Type":"ContainerDied","Data":"e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1"} Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.961623 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6f46bac-b868-4767-8efb-5fdd89267c66","Type":"ContainerDied","Data":"65e89b9664d83b70315f20ff40b3b363a7a1e4bf87f05b6e87a3b80b7285bb51"} Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.961639 4696 scope.go:117] "RemoveContainer" containerID="e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1" Nov 25 10:58:25 crc kubenswrapper[4696]: I1125 10:58:25.961739 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.006186 4696 scope.go:117] "RemoveContainer" containerID="e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1" Nov 25 10:58:26 crc kubenswrapper[4696]: E1125 10:58:26.011041 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1\": container with ID starting with e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1 not found: ID does not exist" containerID="e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.011114 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1"} err="failed to get container status \"e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1\": rpc error: code = NotFound desc = could not find container \"e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1\": container with ID starting with e52aa2755835255b4c3583151f579f8cc33d15c8130968c74958a5dda4b4b5f1 not found: ID does not exist" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.021437 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.035747 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.065582 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6f46bac-b868-4767-8efb-5fdd89267c66" path="/var/lib/kubelet/pods/c6f46bac-b868-4767-8efb-5fdd89267c66/volumes" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.066437 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfa5f437-b259-4cd6-8044-e3c1aa7f56a1" path="/var/lib/kubelet/pods/dfa5f437-b259-4cd6-8044-e3c1aa7f56a1/volumes" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.068617 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:58:26 crc kubenswrapper[4696]: E1125 10:58:26.069095 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f46bac-b868-4767-8efb-5fdd89267c66" containerName="nova-scheduler-scheduler" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.069192 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f46bac-b868-4767-8efb-5fdd89267c66" containerName="nova-scheduler-scheduler" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.069503 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6f46bac-b868-4767-8efb-5fdd89267c66" containerName="nova-scheduler-scheduler" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.070297 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.070481 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.073246 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.135633 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.141030 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmsx6\" (UniqueName: \"kubernetes.io/projected/97905b5e-c4ba-4712-8ed6-e1f1c228f683-kube-api-access-bmsx6\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.141719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97905b5e-c4ba-4712-8ed6-e1f1c228f683-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.141849 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97905b5e-c4ba-4712-8ed6-e1f1c228f683-config-data\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.242938 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmsx6\" (UniqueName: \"kubernetes.io/projected/97905b5e-c4ba-4712-8ed6-e1f1c228f683-kube-api-access-bmsx6\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.243315 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97905b5e-c4ba-4712-8ed6-e1f1c228f683-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.243335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97905b5e-c4ba-4712-8ed6-e1f1c228f683-config-data\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.249534 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97905b5e-c4ba-4712-8ed6-e1f1c228f683-config-data\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.250236 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97905b5e-c4ba-4712-8ed6-e1f1c228f683-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.258246 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmsx6\" (UniqueName: \"kubernetes.io/projected/97905b5e-c4ba-4712-8ed6-e1f1c228f683-kube-api-access-bmsx6\") pod \"nova-scheduler-0\" (UID: \"97905b5e-c4ba-4712-8ed6-e1f1c228f683\") " pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.395214 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.959652 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:58:26 crc kubenswrapper[4696]: W1125 10:58:26.962750 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97905b5e_c4ba_4712_8ed6_e1f1c228f683.slice/crio-0b985544c2470a291bc340070140a4665de795276985d40f65fea048b8a35447 WatchSource:0}: Error finding container 0b985544c2470a291bc340070140a4665de795276985d40f65fea048b8a35447: Status 404 returned error can't find the container with id 0b985544c2470a291bc340070140a4665de795276985d40f65fea048b8a35447 Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.971973 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8377dca6-a47c-4ffb-9916-f6505eed1e48","Type":"ContainerStarted","Data":"eaeae17e461bef5f71194367f5dc927a53e7fb772199711522d8a869a240454c"} Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.972035 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8377dca6-a47c-4ffb-9916-f6505eed1e48","Type":"ContainerStarted","Data":"cb0ede13bae9aeda412751d909a0dd3b2fbb07ed9ce3fe27623143b72d1d1a3a"} Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.972050 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8377dca6-a47c-4ffb-9916-f6505eed1e48","Type":"ContainerStarted","Data":"75668560a5a2f4bb1de6646e602717237e157737edf8eac06f815cf6aa9114c1"} Nov 25 10:58:26 crc kubenswrapper[4696]: I1125 10:58:26.998185 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.998167133 podStartE2EDuration="1.998167133s" podCreationTimestamp="2025-11-25 10:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:26.987152862 +0000 UTC m=+1303.800769449" watchObservedRunningTime="2025-11-25 10:58:26.998167133 +0000 UTC m=+1303.811783720" Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.529206 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:43038->10.217.0.191:8775: read: connection reset by peer" Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.529207 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:43036->10.217.0.191:8775: read: connection reset by peer" Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.993064 4696 generic.go:334] "Generic (PLEG): container finished" podID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerID="a1e21147c6240a548d496dccb577f9e853613f365e7d7265b9d79e18445e106e" exitCode=0 Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.993153 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b67cc31d-f86a-4851-9ac2-4936d79f8273","Type":"ContainerDied","Data":"a1e21147c6240a548d496dccb577f9e853613f365e7d7265b9d79e18445e106e"} Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.993437 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b67cc31d-f86a-4851-9ac2-4936d79f8273","Type":"ContainerDied","Data":"6d0e242c0c170f76d0644b9a7c308231db3c748fb05960f21518d21c58b19df6"} Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.993454 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d0e242c0c170f76d0644b9a7c308231db3c748fb05960f21518d21c58b19df6" Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.995159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"97905b5e-c4ba-4712-8ed6-e1f1c228f683","Type":"ContainerStarted","Data":"e09f83786df1cbe451db208c9eac1be30d88032b692c043c8ea6ee496835f381"} Nov 25 10:58:27 crc kubenswrapper[4696]: I1125 10:58:27.995193 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"97905b5e-c4ba-4712-8ed6-e1f1c228f683","Type":"ContainerStarted","Data":"0b985544c2470a291bc340070140a4665de795276985d40f65fea048b8a35447"} Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.014428 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.040933 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.040910806 podStartE2EDuration="3.040910806s" podCreationTimestamp="2025-11-25 10:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:28.012266123 +0000 UTC m=+1304.825882720" watchObservedRunningTime="2025-11-25 10:58:28.040910806 +0000 UTC m=+1304.854527393" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.077240 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67cc31d-f86a-4851-9ac2-4936d79f8273-logs\") pod \"b67cc31d-f86a-4851-9ac2-4936d79f8273\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.077568 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lh5x\" (UniqueName: \"kubernetes.io/projected/b67cc31d-f86a-4851-9ac2-4936d79f8273-kube-api-access-5lh5x\") pod \"b67cc31d-f86a-4851-9ac2-4936d79f8273\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.077714 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-combined-ca-bundle\") pod \"b67cc31d-f86a-4851-9ac2-4936d79f8273\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.077772 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b67cc31d-f86a-4851-9ac2-4936d79f8273-logs" (OuterVolumeSpecName: "logs") pod "b67cc31d-f86a-4851-9ac2-4936d79f8273" (UID: "b67cc31d-f86a-4851-9ac2-4936d79f8273"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.078142 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-config-data\") pod \"b67cc31d-f86a-4851-9ac2-4936d79f8273\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.078293 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-nova-metadata-tls-certs\") pod \"b67cc31d-f86a-4851-9ac2-4936d79f8273\" (UID: \"b67cc31d-f86a-4851-9ac2-4936d79f8273\") " Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.078978 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67cc31d-f86a-4851-9ac2-4936d79f8273-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.103093 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67cc31d-f86a-4851-9ac2-4936d79f8273-kube-api-access-5lh5x" (OuterVolumeSpecName: "kube-api-access-5lh5x") pod "b67cc31d-f86a-4851-9ac2-4936d79f8273" (UID: "b67cc31d-f86a-4851-9ac2-4936d79f8273"). InnerVolumeSpecName "kube-api-access-5lh5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.137430 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b67cc31d-f86a-4851-9ac2-4936d79f8273" (UID: "b67cc31d-f86a-4851-9ac2-4936d79f8273"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.142855 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-config-data" (OuterVolumeSpecName: "config-data") pod "b67cc31d-f86a-4851-9ac2-4936d79f8273" (UID: "b67cc31d-f86a-4851-9ac2-4936d79f8273"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.147512 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b67cc31d-f86a-4851-9ac2-4936d79f8273" (UID: "b67cc31d-f86a-4851-9ac2-4936d79f8273"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.181356 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.181396 4696 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.181409 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lh5x\" (UniqueName: \"kubernetes.io/projected/b67cc31d-f86a-4851-9ac2-4936d79f8273-kube-api-access-5lh5x\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:28 crc kubenswrapper[4696]: I1125 10:58:28.181423 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67cc31d-f86a-4851-9ac2-4936d79f8273-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.003528 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.040903 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.049563 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.070517 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:58:29 crc kubenswrapper[4696]: E1125 10:58:29.070916 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-log" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.070945 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-log" Nov 25 10:58:29 crc kubenswrapper[4696]: E1125 10:58:29.070984 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-metadata" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.070990 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-metadata" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.071186 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-log" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.071228 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" containerName="nova-metadata-metadata" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.072194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.078130 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.078220 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.087562 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.199862 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-config-data\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.200190 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv5xn\" (UniqueName: \"kubernetes.io/projected/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-kube-api-access-dv5xn\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.200319 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.200362 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-logs\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.200461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.303013 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv5xn\" (UniqueName: \"kubernetes.io/projected/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-kube-api-access-dv5xn\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.303111 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.303142 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-logs\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.303209 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.303303 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-config-data\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.305480 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-logs\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.310471 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-config-data\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.310568 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.318433 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.331030 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv5xn\" (UniqueName: \"kubernetes.io/projected/d808ab02-7a75-4441-a5d0-b92fcaa1eb06-kube-api-access-dv5xn\") pod \"nova-metadata-0\" (UID: \"d808ab02-7a75-4441-a5d0-b92fcaa1eb06\") " pod="openstack/nova-metadata-0" Nov 25 10:58:29 crc kubenswrapper[4696]: I1125 10:58:29.391497 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:58:30 crc kubenswrapper[4696]: I1125 10:58:30.053482 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67cc31d-f86a-4851-9ac2-4936d79f8273" path="/var/lib/kubelet/pods/b67cc31d-f86a-4851-9ac2-4936d79f8273/volumes" Nov 25 10:58:30 crc kubenswrapper[4696]: I1125 10:58:30.442592 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:58:30 crc kubenswrapper[4696]: W1125 10:58:30.448358 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd808ab02_7a75_4441_a5d0_b92fcaa1eb06.slice/crio-b22c6ff58a61caea5144c08642889f72c9f632786830205dbed92a5ad2b4234e WatchSource:0}: Error finding container b22c6ff58a61caea5144c08642889f72c9f632786830205dbed92a5ad2b4234e: Status 404 returned error can't find the container with id b22c6ff58a61caea5144c08642889f72c9f632786830205dbed92a5ad2b4234e Nov 25 10:58:31 crc kubenswrapper[4696]: I1125 10:58:31.042917 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d808ab02-7a75-4441-a5d0-b92fcaa1eb06","Type":"ContainerStarted","Data":"6b7764bfc1f9a9c7f828097e071df2e26720c7c1e8ccab925689a99c1c688e4a"} Nov 25 10:58:31 crc kubenswrapper[4696]: I1125 10:58:31.043209 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d808ab02-7a75-4441-a5d0-b92fcaa1eb06","Type":"ContainerStarted","Data":"dc317253fcf2327bf33e8bcae5a42e17ab6911afced23be92fb919d898afe578"} Nov 25 10:58:31 crc kubenswrapper[4696]: I1125 10:58:31.043223 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d808ab02-7a75-4441-a5d0-b92fcaa1eb06","Type":"ContainerStarted","Data":"b22c6ff58a61caea5144c08642889f72c9f632786830205dbed92a5ad2b4234e"} Nov 25 10:58:31 crc kubenswrapper[4696]: I1125 10:58:31.396268 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:58:34 crc kubenswrapper[4696]: I1125 10:58:34.391880 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:58:34 crc kubenswrapper[4696]: I1125 10:58:34.392241 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:58:35 crc kubenswrapper[4696]: I1125 10:58:35.679136 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:58:35 crc kubenswrapper[4696]: I1125 10:58:35.680156 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:58:36 crc kubenswrapper[4696]: I1125 10:58:36.395795 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:58:36 crc kubenswrapper[4696]: I1125 10:58:36.420588 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:58:36 crc kubenswrapper[4696]: I1125 10:58:36.437015 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=7.436996519 podStartE2EDuration="7.436996519s" podCreationTimestamp="2025-11-25 10:58:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:31.083962751 +0000 UTC m=+1307.897579338" watchObservedRunningTime="2025-11-25 10:58:36.436996519 +0000 UTC m=+1313.250613106" Nov 25 10:58:36 crc kubenswrapper[4696]: I1125 10:58:36.690861 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8377dca6-a47c-4ffb-9916-f6505eed1e48" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:58:36 crc kubenswrapper[4696]: I1125 10:58:36.690903 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8377dca6-a47c-4ffb-9916-f6505eed1e48" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:58:37 crc kubenswrapper[4696]: I1125 10:58:37.132018 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:58:39 crc kubenswrapper[4696]: I1125 10:58:39.391722 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:58:39 crc kubenswrapper[4696]: I1125 10:58:39.392073 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:58:40 crc kubenswrapper[4696]: I1125 10:58:40.411881 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d808ab02-7a75-4441-a5d0-b92fcaa1eb06" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:58:40 crc kubenswrapper[4696]: I1125 10:58:40.411917 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d808ab02-7a75-4441-a5d0-b92fcaa1eb06" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:58:45 crc kubenswrapper[4696]: I1125 10:58:45.225701 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:58:45 crc kubenswrapper[4696]: I1125 10:58:45.684842 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:58:45 crc kubenswrapper[4696]: I1125 10:58:45.685169 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:58:45 crc kubenswrapper[4696]: I1125 10:58:45.686872 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:58:45 crc kubenswrapper[4696]: I1125 10:58:45.692198 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:58:46 crc kubenswrapper[4696]: I1125 10:58:46.188047 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:58:46 crc kubenswrapper[4696]: I1125 10:58:46.471133 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:58:49 crc kubenswrapper[4696]: I1125 10:58:49.398864 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:58:49 crc kubenswrapper[4696]: I1125 10:58:49.400149 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:58:49 crc kubenswrapper[4696]: I1125 10:58:49.407898 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:58:50 crc kubenswrapper[4696]: I1125 10:58:50.243018 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:58:58 crc kubenswrapper[4696]: I1125 10:58:58.965914 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:58:59 crc kubenswrapper[4696]: I1125 10:58:59.765464 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:59:03 crc kubenswrapper[4696]: I1125 10:59:03.732705 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="rabbitmq" containerID="cri-o://9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6" gracePeriod=604796 Nov 25 10:59:03 crc kubenswrapper[4696]: I1125 10:59:03.969827 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="rabbitmq" containerID="cri-o://6b68ac85afb11c275db7068ddade9c64bd80c7ef050364e5e752636e7879b7db" gracePeriod=604796 Nov 25 10:59:05 crc kubenswrapper[4696]: I1125 10:59:05.822655 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Nov 25 10:59:05 crc kubenswrapper[4696]: I1125 10:59:05.987445 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.362718 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.456254 4696 generic.go:334] "Generic (PLEG): container finished" podID="01da0ce5-e289-4f9d-95d2-922e92053827" containerID="9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6" exitCode=0 Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.456343 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01da0ce5-e289-4f9d-95d2-922e92053827","Type":"ContainerDied","Data":"9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6"} Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.456378 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01da0ce5-e289-4f9d-95d2-922e92053827","Type":"ContainerDied","Data":"e646c4f0740ba9eaea56fad2b513d8d83d8b36fab0be6a5bab5d49bfc75650a2"} Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.456400 4696 scope.go:117] "RemoveContainer" containerID="9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.456558 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.460033 4696 generic.go:334] "Generic (PLEG): container finished" podID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerID="6b68ac85afb11c275db7068ddade9c64bd80c7ef050364e5e752636e7879b7db" exitCode=0 Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.460077 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2ff36659-0a3a-4a6f-9f79-414918df363a","Type":"ContainerDied","Data":"6b68ac85afb11c275db7068ddade9c64bd80c7ef050364e5e752636e7879b7db"} Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499537 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-server-conf\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499576 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01da0ce5-e289-4f9d-95d2-922e92053827-pod-info\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499617 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-plugins\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499682 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-erlang-cookie\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499747 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-config-data\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499771 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-tls\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499790 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh895\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-kube-api-access-bh895\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499822 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499874 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01da0ce5-e289-4f9d-95d2-922e92053827-erlang-cookie-secret\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499909 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-confd\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.499954 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-plugins-conf\") pod \"01da0ce5-e289-4f9d-95d2-922e92053827\" (UID: \"01da0ce5-e289-4f9d-95d2-922e92053827\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.503760 4696 scope.go:117] "RemoveContainer" containerID="c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.504419 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.508441 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.510423 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.538463 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01da0ce5-e289-4f9d-95d2-922e92053827-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.543150 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.545965 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.545999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-kube-api-access-bh895" (OuterVolumeSpecName: "kube-api-access-bh895") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "kube-api-access-bh895". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.557440 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-config-data" (OuterVolumeSpecName: "config-data") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.561877 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/01da0ce5-e289-4f9d-95d2-922e92053827-pod-info" (OuterVolumeSpecName: "pod-info") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.585635 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-server-conf" (OuterVolumeSpecName: "server-conf") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601795 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601913 4696 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601933 4696 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601941 4696 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01da0ce5-e289-4f9d-95d2-922e92053827-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601949 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601959 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601967 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01da0ce5-e289-4f9d-95d2-922e92053827-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601975 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.601984 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh895\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-kube-api-access-bh895\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.602013 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.602023 4696 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01da0ce5-e289-4f9d-95d2-922e92053827-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.606983 4696 scope.go:117] "RemoveContainer" containerID="9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6" Nov 25 10:59:10 crc kubenswrapper[4696]: E1125 10:59:10.607693 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6\": container with ID starting with 9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6 not found: ID does not exist" containerID="9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.607750 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6"} err="failed to get container status \"9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6\": rpc error: code = NotFound desc = could not find container \"9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6\": container with ID starting with 9578f86b7e0715942a9f18aa183ecf28c1fbc5e0aa92774d4b2550cc5b219ae6 not found: ID does not exist" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.607773 4696 scope.go:117] "RemoveContainer" containerID="c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b" Nov 25 10:59:10 crc kubenswrapper[4696]: E1125 10:59:10.608357 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b\": container with ID starting with c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b not found: ID does not exist" containerID="c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.608408 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b"} err="failed to get container status \"c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b\": rpc error: code = NotFound desc = could not find container \"c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b\": container with ID starting with c8f473dc09ef51cdff40ee83a774532ea3871776fe4205e51bc86c614a05a35b not found: ID does not exist" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.654954 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.694990 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "01da0ce5-e289-4f9d-95d2-922e92053827" (UID: "01da0ce5-e289-4f9d-95d2-922e92053827"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704157 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-erlang-cookie\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704220 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704304 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-plugins-conf\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704361 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-confd\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704418 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ff36659-0a3a-4a6f-9f79-414918df363a-erlang-cookie-secret\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704444 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-plugins\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704510 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ff36659-0a3a-4a6f-9f79-414918df363a-pod-info\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704553 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-server-conf\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704593 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-config-data\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704631 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzm8j\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-kube-api-access-lzm8j\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.704697 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-tls\") pod \"2ff36659-0a3a-4a6f-9f79-414918df363a\" (UID: \"2ff36659-0a3a-4a6f-9f79-414918df363a\") " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.705236 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.705262 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01da0ce5-e289-4f9d-95d2-922e92053827-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.710494 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.712149 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.713347 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.713351 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.715250 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.716794 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff36659-0a3a-4a6f-9f79-414918df363a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.718516 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-kube-api-access-lzm8j" (OuterVolumeSpecName: "kube-api-access-lzm8j") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "kube-api-access-lzm8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.720267 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2ff36659-0a3a-4a6f-9f79-414918df363a-pod-info" (OuterVolumeSpecName: "pod-info") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.779360 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-config-data" (OuterVolumeSpecName: "config-data") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.801281 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.801478 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-server-conf" (OuterVolumeSpecName: "server-conf") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806636 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzm8j\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-kube-api-access-lzm8j\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806694 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806708 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806741 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806755 4696 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806766 4696 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ff36659-0a3a-4a6f-9f79-414918df363a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806777 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806787 4696 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ff36659-0a3a-4a6f-9f79-414918df363a-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806798 4696 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.806808 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ff36659-0a3a-4a6f-9f79-414918df363a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.811766 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.828183 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:59:10 crc kubenswrapper[4696]: E1125 10:59:10.828833 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="setup-container" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.828921 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="setup-container" Nov 25 10:59:10 crc kubenswrapper[4696]: E1125 10:59:10.828996 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="rabbitmq" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.829093 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="rabbitmq" Nov 25 10:59:10 crc kubenswrapper[4696]: E1125 10:59:10.829195 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="setup-container" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.829266 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="setup-container" Nov 25 10:59:10 crc kubenswrapper[4696]: E1125 10:59:10.829348 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="rabbitmq" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.829415 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="rabbitmq" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.829723 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" containerName="rabbitmq" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.829830 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" containerName="rabbitmq" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.832752 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.836289 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.836433 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-j2rpp" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.836534 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.836639 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.842906 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.846388 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.850483 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.854076 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.856713 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.883748 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2ff36659-0a3a-4a6f-9f79-414918df363a" (UID: "2ff36659-0a3a-4a6f-9f79-414918df363a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.910568 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b161351-5bdf-4cfe-895b-29381797d908-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.910646 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.910760 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.910890 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.910952 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.911015 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.911057 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.911098 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-config-data\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.911125 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b161351-5bdf-4cfe-895b-29381797d908-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.911171 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.911212 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d67cs\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-kube-api-access-d67cs\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.912453 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:10 crc kubenswrapper[4696]: I1125 10:59:10.912479 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ff36659-0a3a-4a6f-9f79-414918df363a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014427 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014780 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014809 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014837 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b161351-5bdf-4cfe-895b-29381797d908-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014856 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-config-data\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014888 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014928 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d67cs\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-kube-api-access-d67cs\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.014996 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b161351-5bdf-4cfe-895b-29381797d908-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.015020 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.015079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.016584 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.017112 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-config-data\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.017258 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.018506 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.018533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.019200 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b161351-5bdf-4cfe-895b-29381797d908-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.019468 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.021544 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b161351-5bdf-4cfe-895b-29381797d908-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.021881 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.021966 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b161351-5bdf-4cfe-895b-29381797d908-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.031955 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d67cs\" (UniqueName: \"kubernetes.io/projected/9b161351-5bdf-4cfe-895b-29381797d908-kube-api-access-d67cs\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.057736 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9b161351-5bdf-4cfe-895b-29381797d908\") " pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.158507 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.469979 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2ff36659-0a3a-4a6f-9f79-414918df363a","Type":"ContainerDied","Data":"64019ebf7a097dc66556f3c0d58c6e0e25bdb8e664ccab550551500e85eb5b29"} Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.470285 4696 scope.go:117] "RemoveContainer" containerID="6b68ac85afb11c275db7068ddade9c64bd80c7ef050364e5e752636e7879b7db" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.470089 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.493345 4696 scope.go:117] "RemoveContainer" containerID="fa9cbf49307a6d897269a72edeb254cafcbb7dab2a48856998a38a4b27815fba" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.514543 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.545725 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.566735 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.568690 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.572116 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.572222 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.572277 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.572340 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.572461 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g5b2j" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.572475 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.579182 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.580312 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.625777 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.625821 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf5b7831-caa9-4411-9386-cf99ba7f8a20-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626011 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46tsx\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-kube-api-access-46tsx\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626194 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626310 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626357 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626426 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626470 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf5b7831-caa9-4411-9386-cf99ba7f8a20-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626507 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.626586 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.637222 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.727829 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.727869 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf5b7831-caa9-4411-9386-cf99ba7f8a20-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.727897 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.727933 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.727978 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.727997 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf5b7831-caa9-4411-9386-cf99ba7f8a20-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728037 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46tsx\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-kube-api-access-46tsx\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728057 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728092 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728128 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728342 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.728431 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.729278 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.729322 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.729870 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf5b7831-caa9-4411-9386-cf99ba7f8a20-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.730705 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.731581 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.734049 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf5b7831-caa9-4411-9386-cf99ba7f8a20-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.740474 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.744705 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf5b7831-caa9-4411-9386-cf99ba7f8a20-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.772858 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-lp78f"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.790601 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.802230 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.803496 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.803971 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46tsx\" (UniqueName: \"kubernetes.io/projected/bf5b7831-caa9-4411-9386-cf99ba7f8a20-kube-api-access-46tsx\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf5b7831-caa9-4411-9386-cf99ba7f8a20\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.815024 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-lp78f"] Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832164 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-config\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832268 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-svc\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832367 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832429 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x9cj\" (UniqueName: \"kubernetes.io/projected/ff430793-3355-4ad6-9ec7-d7b365531d19-kube-api-access-6x9cj\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832475 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832495 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.832541 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.901844 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.933884 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x9cj\" (UniqueName: \"kubernetes.io/projected/ff430793-3355-4ad6-9ec7-d7b365531d19-kube-api-access-6x9cj\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.933944 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.933964 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.934000 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.934034 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-config\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.934063 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-svc\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.934126 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.935377 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-svc\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.935378 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-config\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.935750 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.935836 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.936486 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.936501 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:11 crc kubenswrapper[4696]: I1125 10:59:11.950795 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x9cj\" (UniqueName: \"kubernetes.io/projected/ff430793-3355-4ad6-9ec7-d7b365531d19-kube-api-access-6x9cj\") pod \"dnsmasq-dns-d558885bc-lp78f\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.066532 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01da0ce5-e289-4f9d-95d2-922e92053827" path="/var/lib/kubelet/pods/01da0ce5-e289-4f9d-95d2-922e92053827/volumes" Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.068512 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff36659-0a3a-4a6f-9f79-414918df363a" path="/var/lib/kubelet/pods/2ff36659-0a3a-4a6f-9f79-414918df363a/volumes" Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.125165 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:12 crc kubenswrapper[4696]: W1125 10:59:12.429179 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf5b7831_caa9_4411_9386_cf99ba7f8a20.slice/crio-81cd6991f84cf2bcfd0a13b6126d82a3be729e1e498680a14f37495f7f164f85 WatchSource:0}: Error finding container 81cd6991f84cf2bcfd0a13b6126d82a3be729e1e498680a14f37495f7f164f85: Status 404 returned error can't find the container with id 81cd6991f84cf2bcfd0a13b6126d82a3be729e1e498680a14f37495f7f164f85 Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.432698 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.482445 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf5b7831-caa9-4411-9386-cf99ba7f8a20","Type":"ContainerStarted","Data":"81cd6991f84cf2bcfd0a13b6126d82a3be729e1e498680a14f37495f7f164f85"} Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.484681 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b161351-5bdf-4cfe-895b-29381797d908","Type":"ContainerStarted","Data":"b68775110e15565c9a3dc21036bcc7739cb89eae60fb3bcf5c4cfb3b7b4312c5"} Nov 25 10:59:12 crc kubenswrapper[4696]: I1125 10:59:12.623871 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-lp78f"] Nov 25 10:59:12 crc kubenswrapper[4696]: W1125 10:59:12.624653 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff430793_3355_4ad6_9ec7_d7b365531d19.slice/crio-852e484e1e18a7163451383465005a079ebd86a5c089707d4f95a24bcd30e8c8 WatchSource:0}: Error finding container 852e484e1e18a7163451383465005a079ebd86a5c089707d4f95a24bcd30e8c8: Status 404 returned error can't find the container with id 852e484e1e18a7163451383465005a079ebd86a5c089707d4f95a24bcd30e8c8 Nov 25 10:59:13 crc kubenswrapper[4696]: I1125 10:59:13.494209 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerID="682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5" exitCode=0 Nov 25 10:59:13 crc kubenswrapper[4696]: I1125 10:59:13.494254 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-lp78f" event={"ID":"ff430793-3355-4ad6-9ec7-d7b365531d19","Type":"ContainerDied","Data":"682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5"} Nov 25 10:59:13 crc kubenswrapper[4696]: I1125 10:59:13.494490 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-lp78f" event={"ID":"ff430793-3355-4ad6-9ec7-d7b365531d19","Type":"ContainerStarted","Data":"852e484e1e18a7163451383465005a079ebd86a5c089707d4f95a24bcd30e8c8"} Nov 25 10:59:13 crc kubenswrapper[4696]: I1125 10:59:13.498819 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b161351-5bdf-4cfe-895b-29381797d908","Type":"ContainerStarted","Data":"01bf74afd70d07719590023612ba6ffca4e4a88ddec17aacc6fc05aa914a20b7"} Nov 25 10:59:14 crc kubenswrapper[4696]: I1125 10:59:14.517346 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf5b7831-caa9-4411-9386-cf99ba7f8a20","Type":"ContainerStarted","Data":"59dfa7f55368f0c26db9d4cc80d0e6b88411972cc37f8da446b6adca771c50ae"} Nov 25 10:59:14 crc kubenswrapper[4696]: I1125 10:59:14.530395 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-lp78f" event={"ID":"ff430793-3355-4ad6-9ec7-d7b365531d19","Type":"ContainerStarted","Data":"d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d"} Nov 25 10:59:14 crc kubenswrapper[4696]: I1125 10:59:14.530454 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:14 crc kubenswrapper[4696]: I1125 10:59:14.579597 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-lp78f" podStartSLOduration=3.579581662 podStartE2EDuration="3.579581662s" podCreationTimestamp="2025-11-25 10:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:14.572899879 +0000 UTC m=+1351.386516496" watchObservedRunningTime="2025-11-25 10:59:14.579581662 +0000 UTC m=+1351.393198249" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.126550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.186606 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-kbdbq"] Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.186877 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerName="dnsmasq-dns" containerID="cri-o://8525bc7bbac5dc4f6924ad859e4b8311e5a96782b323e57bd3f56f9e34cf1404" gracePeriod=10 Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.455464 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-qdbhs"] Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.462821 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.504240 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-qdbhs"] Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.593198 4696 generic.go:334] "Generic (PLEG): container finished" podID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerID="8525bc7bbac5dc4f6924ad859e4b8311e5a96782b323e57bd3f56f9e34cf1404" exitCode=0 Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.593244 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" event={"ID":"bb61a07b-9c8f-4f67-b714-05d82b3b1715","Type":"ContainerDied","Data":"8525bc7bbac5dc4f6924ad859e4b8311e5a96782b323e57bd3f56f9e34cf1404"} Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651416 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651656 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651726 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mvch\" (UniqueName: \"kubernetes.io/projected/6b74d08c-9100-49d2-8483-34d4c76c0c6b-kube-api-access-6mvch\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651754 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651792 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651869 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.651981 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-config\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.755686 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-config\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756091 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756149 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756169 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mvch\" (UniqueName: \"kubernetes.io/projected/6b74d08c-9100-49d2-8483-34d4c76c0c6b-kube-api-access-6mvch\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756186 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756203 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756220 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.756950 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-config\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.757549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.758139 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.758336 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.758790 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.758888 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6b74d08c-9100-49d2-8483-34d4c76c0c6b-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:22 crc kubenswrapper[4696]: I1125 10:59:22.784203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mvch\" (UniqueName: \"kubernetes.io/projected/6b74d08c-9100-49d2-8483-34d4c76c0c6b-kube-api-access-6mvch\") pod \"dnsmasq-dns-798f84bf5f-qdbhs\" (UID: \"6b74d08c-9100-49d2-8483-34d4c76c0c6b\") " pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.056597 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.080506 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.162572 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-swift-storage-0\") pod \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.162877 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-nb\") pod \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.162920 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-sb\") pod \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.163019 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znhr6\" (UniqueName: \"kubernetes.io/projected/bb61a07b-9c8f-4f67-b714-05d82b3b1715-kube-api-access-znhr6\") pod \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.163054 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-config\") pod \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.163082 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-svc\") pod \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\" (UID: \"bb61a07b-9c8f-4f67-b714-05d82b3b1715\") " Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.169402 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb61a07b-9c8f-4f67-b714-05d82b3b1715-kube-api-access-znhr6" (OuterVolumeSpecName: "kube-api-access-znhr6") pod "bb61a07b-9c8f-4f67-b714-05d82b3b1715" (UID: "bb61a07b-9c8f-4f67-b714-05d82b3b1715"). InnerVolumeSpecName "kube-api-access-znhr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.242333 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb61a07b-9c8f-4f67-b714-05d82b3b1715" (UID: "bb61a07b-9c8f-4f67-b714-05d82b3b1715"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.250514 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb61a07b-9c8f-4f67-b714-05d82b3b1715" (UID: "bb61a07b-9c8f-4f67-b714-05d82b3b1715"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.258014 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb61a07b-9c8f-4f67-b714-05d82b3b1715" (UID: "bb61a07b-9c8f-4f67-b714-05d82b3b1715"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.259745 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-config" (OuterVolumeSpecName: "config") pod "bb61a07b-9c8f-4f67-b714-05d82b3b1715" (UID: "bb61a07b-9c8f-4f67-b714-05d82b3b1715"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.263528 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb61a07b-9c8f-4f67-b714-05d82b3b1715" (UID: "bb61a07b-9c8f-4f67-b714-05d82b3b1715"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.266602 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.266729 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.266796 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.266857 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znhr6\" (UniqueName: \"kubernetes.io/projected/bb61a07b-9c8f-4f67-b714-05d82b3b1715-kube-api-access-znhr6\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.266943 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.267007 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb61a07b-9c8f-4f67-b714-05d82b3b1715-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.604490 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-qdbhs"] Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.607809 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" event={"ID":"bb61a07b-9c8f-4f67-b714-05d82b3b1715","Type":"ContainerDied","Data":"cb382a9190fd086b53aea6980331b723bed7ee56ec8da81dcb10cddf2b6637cf"} Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.607889 4696 scope.go:117] "RemoveContainer" containerID="8525bc7bbac5dc4f6924ad859e4b8311e5a96782b323e57bd3f56f9e34cf1404" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.608062 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-kbdbq" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.794121 4696 scope.go:117] "RemoveContainer" containerID="2882758aa48311ddca88e2b788cea603b9658e5c7eef2bd0410a41d74f97dd4b" Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.825220 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-kbdbq"] Nov 25 10:59:23 crc kubenswrapper[4696]: I1125 10:59:23.837247 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-kbdbq"] Nov 25 10:59:24 crc kubenswrapper[4696]: I1125 10:59:24.052854 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" path="/var/lib/kubelet/pods/bb61a07b-9c8f-4f67-b714-05d82b3b1715/volumes" Nov 25 10:59:24 crc kubenswrapper[4696]: I1125 10:59:24.621886 4696 generic.go:334] "Generic (PLEG): container finished" podID="6b74d08c-9100-49d2-8483-34d4c76c0c6b" containerID="a929cefd8c17c5a28bcbc203be58e700b302607419548388aa4c75bc19ef6fbb" exitCode=0 Nov 25 10:59:24 crc kubenswrapper[4696]: I1125 10:59:24.621937 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" event={"ID":"6b74d08c-9100-49d2-8483-34d4c76c0c6b","Type":"ContainerDied","Data":"a929cefd8c17c5a28bcbc203be58e700b302607419548388aa4c75bc19ef6fbb"} Nov 25 10:59:24 crc kubenswrapper[4696]: I1125 10:59:24.621968 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" event={"ID":"6b74d08c-9100-49d2-8483-34d4c76c0c6b","Type":"ContainerStarted","Data":"572cc3d54d76539260603980674cf6d90a7ccaa24edcf683e6e90934964e0354"} Nov 25 10:59:25 crc kubenswrapper[4696]: I1125 10:59:25.632867 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" event={"ID":"6b74d08c-9100-49d2-8483-34d4c76c0c6b","Type":"ContainerStarted","Data":"6589728ab6d07c92dd8a1b9c7abf2564a593690ac79fb09d8948845dce6da305"} Nov 25 10:59:25 crc kubenswrapper[4696]: I1125 10:59:25.633305 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:25 crc kubenswrapper[4696]: I1125 10:59:25.657350 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" podStartSLOduration=3.657328718 podStartE2EDuration="3.657328718s" podCreationTimestamp="2025-11-25 10:59:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:25.649968924 +0000 UTC m=+1362.463585521" watchObservedRunningTime="2025-11-25 10:59:25.657328718 +0000 UTC m=+1362.470945305" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.083075 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-798f84bf5f-qdbhs" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.156816 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-lp78f"] Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.157066 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-lp78f" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerName="dnsmasq-dns" containerID="cri-o://d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d" gracePeriod=10 Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.664521 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.706773 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerID="d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d" exitCode=0 Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.706835 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-lp78f" event={"ID":"ff430793-3355-4ad6-9ec7-d7b365531d19","Type":"ContainerDied","Data":"d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d"} Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.706858 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-lp78f" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.706887 4696 scope.go:117] "RemoveContainer" containerID="d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.706869 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-lp78f" event={"ID":"ff430793-3355-4ad6-9ec7-d7b365531d19","Type":"ContainerDied","Data":"852e484e1e18a7163451383465005a079ebd86a5c089707d4f95a24bcd30e8c8"} Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.733272 4696 scope.go:117] "RemoveContainer" containerID="682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.758501 4696 scope.go:117] "RemoveContainer" containerID="d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d" Nov 25 10:59:33 crc kubenswrapper[4696]: E1125 10:59:33.758964 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d\": container with ID starting with d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d not found: ID does not exist" containerID="d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.759005 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d"} err="failed to get container status \"d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d\": rpc error: code = NotFound desc = could not find container \"d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d\": container with ID starting with d31703d53c54d20642357a26ef83d3294c4c0c14925dd4eb0815bf88c827b36d not found: ID does not exist" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.759037 4696 scope.go:117] "RemoveContainer" containerID="682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5" Nov 25 10:59:33 crc kubenswrapper[4696]: E1125 10:59:33.759339 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5\": container with ID starting with 682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5 not found: ID does not exist" containerID="682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.759377 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5"} err="failed to get container status \"682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5\": rpc error: code = NotFound desc = could not find container \"682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5\": container with ID starting with 682de0736a2701ec94f67a300e0228fc944df1174b4546c7f861e1cdce1fa4c5 not found: ID does not exist" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867519 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x9cj\" (UniqueName: \"kubernetes.io/projected/ff430793-3355-4ad6-9ec7-d7b365531d19-kube-api-access-6x9cj\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867590 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-svc\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867611 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-openstack-edpm-ipam\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867657 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-nb\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867724 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-sb\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867778 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-config\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.867838 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-swift-storage-0\") pod \"ff430793-3355-4ad6-9ec7-d7b365531d19\" (UID: \"ff430793-3355-4ad6-9ec7-d7b365531d19\") " Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.893917 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff430793-3355-4ad6-9ec7-d7b365531d19-kube-api-access-6x9cj" (OuterVolumeSpecName: "kube-api-access-6x9cj") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "kube-api-access-6x9cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:33 crc kubenswrapper[4696]: I1125 10:59:33.970551 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x9cj\" (UniqueName: \"kubernetes.io/projected/ff430793-3355-4ad6-9ec7-d7b365531d19-kube-api-access-6x9cj\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.026302 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.074135 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-config" (OuterVolumeSpecName: "config") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.074548 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.074577 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.089191 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.098873 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.103106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.109032 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ff430793-3355-4ad6-9ec7-d7b365531d19" (UID: "ff430793-3355-4ad6-9ec7-d7b365531d19"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.176011 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.176234 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.176294 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.176354 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff430793-3355-4ad6-9ec7-d7b365531d19-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.367530 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-lp78f"] Nov 25 10:59:34 crc kubenswrapper[4696]: I1125 10:59:34.378022 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-lp78f"] Nov 25 10:59:36 crc kubenswrapper[4696]: I1125 10:59:36.058378 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" path="/var/lib/kubelet/pods/ff430793-3355-4ad6-9ec7-d7b365531d19/volumes" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.836753 4696 generic.go:334] "Generic (PLEG): container finished" podID="bf5b7831-caa9-4411-9386-cf99ba7f8a20" containerID="59dfa7f55368f0c26db9d4cc80d0e6b88411972cc37f8da446b6adca771c50ae" exitCode=0 Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.836841 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf5b7831-caa9-4411-9386-cf99ba7f8a20","Type":"ContainerDied","Data":"59dfa7f55368f0c26db9d4cc80d0e6b88411972cc37f8da446b6adca771c50ae"} Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.839228 4696 generic.go:334] "Generic (PLEG): container finished" podID="9b161351-5bdf-4cfe-895b-29381797d908" containerID="01bf74afd70d07719590023612ba6ffca4e4a88ddec17aacc6fc05aa914a20b7" exitCode=0 Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.839257 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b161351-5bdf-4cfe-895b-29381797d908","Type":"ContainerDied","Data":"01bf74afd70d07719590023612ba6ffca4e4a88ddec17aacc6fc05aa914a20b7"} Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.985650 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svx82"] Nov 25 10:59:45 crc kubenswrapper[4696]: E1125 10:59:45.987180 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerName="init" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.987199 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerName="init" Nov 25 10:59:45 crc kubenswrapper[4696]: E1125 10:59:45.987215 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerName="dnsmasq-dns" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.987221 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerName="dnsmasq-dns" Nov 25 10:59:45 crc kubenswrapper[4696]: E1125 10:59:45.987273 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerName="dnsmasq-dns" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.987281 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerName="dnsmasq-dns" Nov 25 10:59:45 crc kubenswrapper[4696]: E1125 10:59:45.987351 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerName="init" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.987359 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerName="init" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.987804 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff430793-3355-4ad6-9ec7-d7b365531d19" containerName="dnsmasq-dns" Nov 25 10:59:45 crc kubenswrapper[4696]: I1125 10:59:45.987865 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb61a07b-9c8f-4f67-b714-05d82b3b1715" containerName="dnsmasq-dns" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.000572 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.061324 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svx82"] Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.104189 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dm6hs"] Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.106090 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.117135 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dm6hs"] Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.130195 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csdpr\" (UniqueName: \"kubernetes.io/projected/f7035564-90f9-4dae-ac7a-475b8782825b-kube-api-access-csdpr\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.130279 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-utilities\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.130433 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-catalog-content\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.231987 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-utilities\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.232041 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-catalog-content\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.232117 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwc94\" (UniqueName: \"kubernetes.io/projected/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-kube-api-access-jwc94\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.232154 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csdpr\" (UniqueName: \"kubernetes.io/projected/f7035564-90f9-4dae-ac7a-475b8782825b-kube-api-access-csdpr\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.232190 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-utilities\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.232283 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-catalog-content\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.232843 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-catalog-content\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.233158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-utilities\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.255481 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csdpr\" (UniqueName: \"kubernetes.io/projected/f7035564-90f9-4dae-ac7a-475b8782825b-kube-api-access-csdpr\") pod \"community-operators-svx82\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.333675 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwc94\" (UniqueName: \"kubernetes.io/projected/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-kube-api-access-jwc94\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.333830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-utilities\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.333851 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-catalog-content\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.334245 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-catalog-content\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.334311 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-utilities\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.354152 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwc94\" (UniqueName: \"kubernetes.io/projected/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-kube-api-access-jwc94\") pod \"certified-operators-dm6hs\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.445897 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svx82" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.488931 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.853091 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf5b7831-caa9-4411-9386-cf99ba7f8a20","Type":"ContainerStarted","Data":"a15997efbad44f9823cb1a02effe11fa0d7ab1b2bfd18c89a6bd4a608fe7409e"} Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.853554 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.856901 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b161351-5bdf-4cfe-895b-29381797d908","Type":"ContainerStarted","Data":"a05c3080f8e4f172424c85cb9d5a20b92d6e06de4080cbab626840c53a872f3c"} Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.857384 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.929206 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.92862637 podStartE2EDuration="35.92862637s" podCreationTimestamp="2025-11-25 10:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:46.917176936 +0000 UTC m=+1383.730793523" watchObservedRunningTime="2025-11-25 10:59:46.92862637 +0000 UTC m=+1383.742242957" Nov 25 10:59:46 crc kubenswrapper[4696]: I1125 10:59:46.982849 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.982443733 podStartE2EDuration="36.982443733s" podCreationTimestamp="2025-11-25 10:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:46.957825057 +0000 UTC m=+1383.771441634" watchObservedRunningTime="2025-11-25 10:59:46.982443733 +0000 UTC m=+1383.796060340" Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.055721 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svx82"] Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.173533 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dm6hs"] Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.866953 4696 generic.go:334] "Generic (PLEG): container finished" podID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerID="2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de" exitCode=0 Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.867046 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerDied","Data":"2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de"} Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.868861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerStarted","Data":"d695344da4838e1aae94f483d517e302150964a8e42136c1b673edaa5c54ba87"} Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.871092 4696 generic.go:334] "Generic (PLEG): container finished" podID="f7035564-90f9-4dae-ac7a-475b8782825b" containerID="3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938" exitCode=0 Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.871176 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerDied","Data":"3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938"} Nov 25 10:59:47 crc kubenswrapper[4696]: I1125 10:59:47.871211 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerStarted","Data":"3680ef82a867784082d3199b8d92522525f904a7c5ff68d3d3c7509ad3c4b18b"} Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.300964 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qxf7j"] Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.303022 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.313996 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxf7j"] Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.389362 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmm8t\" (UniqueName: \"kubernetes.io/projected/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-kube-api-access-pmm8t\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.389427 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-catalog-content\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.389465 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-utilities\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.491753 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmm8t\" (UniqueName: \"kubernetes.io/projected/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-kube-api-access-pmm8t\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.491816 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-catalog-content\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.491855 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-utilities\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.492394 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-utilities\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.492969 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-catalog-content\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.494307 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6xg6h"] Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.496451 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.512344 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmm8t\" (UniqueName: \"kubernetes.io/projected/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-kube-api-access-pmm8t\") pod \"redhat-marketplace-qxf7j\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.526253 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xg6h"] Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.593766 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjkfs\" (UniqueName: \"kubernetes.io/projected/2fd52a93-18b2-424b-8a35-d09d44f47d7b-kube-api-access-bjkfs\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.594093 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-catalog-content\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.594219 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-utilities\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.630081 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.699027 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjkfs\" (UniqueName: \"kubernetes.io/projected/2fd52a93-18b2-424b-8a35-d09d44f47d7b-kube-api-access-bjkfs\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.699458 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-catalog-content\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.699547 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-utilities\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.700197 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-utilities\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.700508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-catalog-content\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.795503 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjkfs\" (UniqueName: \"kubernetes.io/projected/2fd52a93-18b2-424b-8a35-d09d44f47d7b-kube-api-access-bjkfs\") pod \"redhat-operators-6xg6h\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:48 crc kubenswrapper[4696]: I1125 10:59:48.865091 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.064715 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxf7j"] Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.299197 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xg6h"] Nov 25 10:59:49 crc kubenswrapper[4696]: W1125 10:59:49.302024 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fd52a93_18b2_424b_8a35_d09d44f47d7b.slice/crio-fea8f5ba69957dc742dd47db8bd4156b1a229567144fda98716669422105bbb0 WatchSource:0}: Error finding container fea8f5ba69957dc742dd47db8bd4156b1a229567144fda98716669422105bbb0: Status 404 returned error can't find the container with id fea8f5ba69957dc742dd47db8bd4156b1a229567144fda98716669422105bbb0 Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.909642 4696 generic.go:334] "Generic (PLEG): container finished" podID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerID="6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1" exitCode=0 Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.909862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerDied","Data":"6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1"} Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.910007 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerStarted","Data":"fea8f5ba69957dc742dd47db8bd4156b1a229567144fda98716669422105bbb0"} Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.912264 4696 generic.go:334] "Generic (PLEG): container finished" podID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerID="de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86" exitCode=0 Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.912328 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerDied","Data":"de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86"} Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.912355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerStarted","Data":"cc834ec7c61acb1e86dc0691f5202ec9f00f659b9a16baf00f5284ddfb5d5f4e"} Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.918165 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerStarted","Data":"ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d"} Nov 25 10:59:49 crc kubenswrapper[4696]: I1125 10:59:49.921709 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerStarted","Data":"1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4"} Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.461498 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx"] Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.463236 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.472625 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.472810 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.472926 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.478048 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.509383 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx"] Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.650342 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.650681 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmf6p\" (UniqueName: \"kubernetes.io/projected/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-kube-api-access-cmf6p\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.650781 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.650858 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.752099 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.752144 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmf6p\" (UniqueName: \"kubernetes.io/projected/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-kube-api-access-cmf6p\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.752376 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.752421 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.765118 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.765541 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.770183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.779486 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmf6p\" (UniqueName: \"kubernetes.io/projected/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-kube-api-access-cmf6p\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:50 crc kubenswrapper[4696]: I1125 10:59:50.781829 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 10:59:51 crc kubenswrapper[4696]: I1125 10:59:50.988345 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerStarted","Data":"da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b"} Nov 25 10:59:52 crc kubenswrapper[4696]: I1125 10:59:52.002471 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerStarted","Data":"120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7"} Nov 25 10:59:52 crc kubenswrapper[4696]: I1125 10:59:52.960238 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx"] Nov 25 10:59:53 crc kubenswrapper[4696]: I1125 10:59:53.037889 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" event={"ID":"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a","Type":"ContainerStarted","Data":"6964730550e50fad669b0263b0291ae5bab5119710a5f01103a7ba4457b4cea5"} Nov 25 10:59:54 crc kubenswrapper[4696]: I1125 10:59:54.048611 4696 generic.go:334] "Generic (PLEG): container finished" podID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerID="ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d" exitCode=0 Nov 25 10:59:54 crc kubenswrapper[4696]: I1125 10:59:54.051869 4696 generic.go:334] "Generic (PLEG): container finished" podID="f7035564-90f9-4dae-ac7a-475b8782825b" containerID="1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4" exitCode=0 Nov 25 10:59:54 crc kubenswrapper[4696]: I1125 10:59:54.054154 4696 generic.go:334] "Generic (PLEG): container finished" podID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerID="120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7" exitCode=0 Nov 25 10:59:54 crc kubenswrapper[4696]: I1125 10:59:54.073488 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerDied","Data":"ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d"} Nov 25 10:59:54 crc kubenswrapper[4696]: I1125 10:59:54.073553 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerDied","Data":"1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4"} Nov 25 10:59:54 crc kubenswrapper[4696]: I1125 10:59:54.073573 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerDied","Data":"120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7"} Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.142689 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerStarted","Data":"80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2"} Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.148228 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerStarted","Data":"2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32"} Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.151911 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerStarted","Data":"65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b"} Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.174096 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svx82" podStartSLOduration=3.616145037 podStartE2EDuration="13.174073507s" podCreationTimestamp="2025-11-25 10:59:45 +0000 UTC" firstStartedPulling="2025-11-25 10:59:47.872721576 +0000 UTC m=+1384.686338153" lastFinishedPulling="2025-11-25 10:59:57.430650036 +0000 UTC m=+1394.244266623" observedRunningTime="2025-11-25 10:59:58.163412893 +0000 UTC m=+1394.977029490" watchObservedRunningTime="2025-11-25 10:59:58.174073507 +0000 UTC m=+1394.987690104" Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.212799 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dm6hs" podStartSLOduration=5.567651289 podStartE2EDuration="12.212777966s" podCreationTimestamp="2025-11-25 10:59:46 +0000 UTC" firstStartedPulling="2025-11-25 10:59:47.868707642 +0000 UTC m=+1384.682324229" lastFinishedPulling="2025-11-25 10:59:54.513834319 +0000 UTC m=+1391.327450906" observedRunningTime="2025-11-25 10:59:58.203351362 +0000 UTC m=+1395.016967969" watchObservedRunningTime="2025-11-25 10:59:58.212777966 +0000 UTC m=+1395.026394563" Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.231429 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qxf7j" podStartSLOduration=5.609878038 podStartE2EDuration="10.231405908s" podCreationTimestamp="2025-11-25 10:59:48 +0000 UTC" firstStartedPulling="2025-11-25 10:59:49.914364451 +0000 UTC m=+1386.727981028" lastFinishedPulling="2025-11-25 10:59:54.535892311 +0000 UTC m=+1391.349508898" observedRunningTime="2025-11-25 10:59:58.225109914 +0000 UTC m=+1395.038726501" watchObservedRunningTime="2025-11-25 10:59:58.231405908 +0000 UTC m=+1395.045022495" Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.631123 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:58 crc kubenswrapper[4696]: I1125 10:59:58.631202 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 10:59:59 crc kubenswrapper[4696]: I1125 10:59:59.167096 4696 generic.go:334] "Generic (PLEG): container finished" podID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerID="da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b" exitCode=0 Nov 25 10:59:59 crc kubenswrapper[4696]: I1125 10:59:59.167272 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerDied","Data":"da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b"} Nov 25 10:59:59 crc kubenswrapper[4696]: I1125 10:59:59.740424 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-qxf7j" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="registry-server" probeResult="failure" output=< Nov 25 10:59:59 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 10:59:59 crc kubenswrapper[4696]: > Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.175803 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl"] Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.177082 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.192041 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.192240 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.228874 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl"] Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.297962 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v69k6\" (UniqueName: \"kubernetes.io/projected/510e1af6-4069-48d0-a02c-17799da5eeed-kube-api-access-v69k6\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.298269 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/510e1af6-4069-48d0-a02c-17799da5eeed-config-volume\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.298382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/510e1af6-4069-48d0-a02c-17799da5eeed-secret-volume\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.400813 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/510e1af6-4069-48d0-a02c-17799da5eeed-config-volume\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.400888 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/510e1af6-4069-48d0-a02c-17799da5eeed-secret-volume\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.401065 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v69k6\" (UniqueName: \"kubernetes.io/projected/510e1af6-4069-48d0-a02c-17799da5eeed-kube-api-access-v69k6\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.402356 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/510e1af6-4069-48d0-a02c-17799da5eeed-config-volume\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.414264 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/510e1af6-4069-48d0-a02c-17799da5eeed-secret-volume\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.454483 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v69k6\" (UniqueName: \"kubernetes.io/projected/510e1af6-4069-48d0-a02c-17799da5eeed-kube-api-access-v69k6\") pod \"collect-profiles-29401140-56scl\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:00 crc kubenswrapper[4696]: I1125 11:00:00.507249 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:01 crc kubenswrapper[4696]: I1125 11:00:01.166882 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9b161351-5bdf-4cfe-895b-29381797d908" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.205:5671: connect: connection refused" Nov 25 11:00:01 crc kubenswrapper[4696]: I1125 11:00:01.909555 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="bf5b7831-caa9-4411-9386-cf99ba7f8a20" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.206:5671: connect: connection refused" Nov 25 11:00:06 crc kubenswrapper[4696]: I1125 11:00:06.447206 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svx82" Nov 25 11:00:06 crc kubenswrapper[4696]: I1125 11:00:06.447640 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svx82" Nov 25 11:00:06 crc kubenswrapper[4696]: I1125 11:00:06.489266 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 11:00:06 crc kubenswrapper[4696]: I1125 11:00:06.489728 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 11:00:07 crc kubenswrapper[4696]: I1125 11:00:07.505344 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-svx82" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:07 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:07 crc kubenswrapper[4696]: > Nov 25 11:00:07 crc kubenswrapper[4696]: I1125 11:00:07.542714 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dm6hs" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:07 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:07 crc kubenswrapper[4696]: > Nov 25 11:00:08 crc kubenswrapper[4696]: E1125 11:00:08.613390 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Nov 25 11:00:08 crc kubenswrapper[4696]: E1125 11:00:08.613604 4696 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 11:00:08 crc kubenswrapper[4696]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Nov 25 11:00:08 crc kubenswrapper[4696]: - hosts: all Nov 25 11:00:08 crc kubenswrapper[4696]: strategy: linear Nov 25 11:00:08 crc kubenswrapper[4696]: tasks: Nov 25 11:00:08 crc kubenswrapper[4696]: - name: Enable podified-repos Nov 25 11:00:08 crc kubenswrapper[4696]: become: true Nov 25 11:00:08 crc kubenswrapper[4696]: ansible.builtin.shell: | Nov 25 11:00:08 crc kubenswrapper[4696]: set -euxo pipefail Nov 25 11:00:08 crc kubenswrapper[4696]: pushd /var/tmp Nov 25 11:00:08 crc kubenswrapper[4696]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Nov 25 11:00:08 crc kubenswrapper[4696]: pushd repo-setup-main Nov 25 11:00:08 crc kubenswrapper[4696]: python3 -m venv ./venv Nov 25 11:00:08 crc kubenswrapper[4696]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Nov 25 11:00:08 crc kubenswrapper[4696]: ./venv/bin/repo-setup current-podified -b antelope Nov 25 11:00:08 crc kubenswrapper[4696]: popd Nov 25 11:00:08 crc kubenswrapper[4696]: rm -rf repo-setup-main Nov 25 11:00:08 crc kubenswrapper[4696]: Nov 25 11:00:08 crc kubenswrapper[4696]: Nov 25 11:00:08 crc kubenswrapper[4696]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Nov 25 11:00:08 crc kubenswrapper[4696]: edpm_override_hosts: openstack-edpm-ipam Nov 25 11:00:08 crc kubenswrapper[4696]: edpm_service_type: repo-setup Nov 25 11:00:08 crc kubenswrapper[4696]: Nov 25 11:00:08 crc kubenswrapper[4696]: Nov 25 11:00:08 crc kubenswrapper[4696]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cmf6p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx_openstack(4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 25 11:00:08 crc kubenswrapper[4696]: > logger="UnhandledError" Nov 25 11:00:08 crc kubenswrapper[4696]: E1125 11:00:08.616836 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" podUID="4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" Nov 25 11:00:09 crc kubenswrapper[4696]: I1125 11:00:09.162376 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl"] Nov 25 11:00:09 crc kubenswrapper[4696]: W1125 11:00:09.173779 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod510e1af6_4069_48d0_a02c_17799da5eeed.slice/crio-3885679d0c3a2c4100021b2e0682a383932239e8e68583039030cbe384b5d3a5 WatchSource:0}: Error finding container 3885679d0c3a2c4100021b2e0682a383932239e8e68583039030cbe384b5d3a5: Status 404 returned error can't find the container with id 3885679d0c3a2c4100021b2e0682a383932239e8e68583039030cbe384b5d3a5 Nov 25 11:00:09 crc kubenswrapper[4696]: I1125 11:00:09.314818 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerStarted","Data":"065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4"} Nov 25 11:00:09 crc kubenswrapper[4696]: I1125 11:00:09.322752 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" event={"ID":"510e1af6-4069-48d0-a02c-17799da5eeed","Type":"ContainerStarted","Data":"3885679d0c3a2c4100021b2e0682a383932239e8e68583039030cbe384b5d3a5"} Nov 25 11:00:09 crc kubenswrapper[4696]: E1125 11:00:09.327455 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" podUID="4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" Nov 25 11:00:09 crc kubenswrapper[4696]: I1125 11:00:09.343113 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6xg6h" podStartSLOduration=2.625658696 podStartE2EDuration="21.343097794s" podCreationTimestamp="2025-11-25 10:59:48 +0000 UTC" firstStartedPulling="2025-11-25 10:59:49.911448527 +0000 UTC m=+1386.725065124" lastFinishedPulling="2025-11-25 11:00:08.628887625 +0000 UTC m=+1405.442504222" observedRunningTime="2025-11-25 11:00:09.339253589 +0000 UTC m=+1406.152870176" watchObservedRunningTime="2025-11-25 11:00:09.343097794 +0000 UTC m=+1406.156714381" Nov 25 11:00:09 crc kubenswrapper[4696]: I1125 11:00:09.709601 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-qxf7j" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:09 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:09 crc kubenswrapper[4696]: > Nov 25 11:00:10 crc kubenswrapper[4696]: I1125 11:00:10.332532 4696 generic.go:334] "Generic (PLEG): container finished" podID="510e1af6-4069-48d0-a02c-17799da5eeed" containerID="26fdd17402683ff0fb6db77b0b89ff28f50fa2234474262337406fe22d85039a" exitCode=0 Nov 25 11:00:10 crc kubenswrapper[4696]: I1125 11:00:10.332676 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" event={"ID":"510e1af6-4069-48d0-a02c-17799da5eeed","Type":"ContainerDied","Data":"26fdd17402683ff0fb6db77b0b89ff28f50fa2234474262337406fe22d85039a"} Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.164471 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.764838 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.868969 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v69k6\" (UniqueName: \"kubernetes.io/projected/510e1af6-4069-48d0-a02c-17799da5eeed-kube-api-access-v69k6\") pod \"510e1af6-4069-48d0-a02c-17799da5eeed\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.869186 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/510e1af6-4069-48d0-a02c-17799da5eeed-config-volume\") pod \"510e1af6-4069-48d0-a02c-17799da5eeed\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.869238 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/510e1af6-4069-48d0-a02c-17799da5eeed-secret-volume\") pod \"510e1af6-4069-48d0-a02c-17799da5eeed\" (UID: \"510e1af6-4069-48d0-a02c-17799da5eeed\") " Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.869771 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510e1af6-4069-48d0-a02c-17799da5eeed-config-volume" (OuterVolumeSpecName: "config-volume") pod "510e1af6-4069-48d0-a02c-17799da5eeed" (UID: "510e1af6-4069-48d0-a02c-17799da5eeed"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.875954 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/510e1af6-4069-48d0-a02c-17799da5eeed-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "510e1af6-4069-48d0-a02c-17799da5eeed" (UID: "510e1af6-4069-48d0-a02c-17799da5eeed"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.876038 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/510e1af6-4069-48d0-a02c-17799da5eeed-kube-api-access-v69k6" (OuterVolumeSpecName: "kube-api-access-v69k6") pod "510e1af6-4069-48d0-a02c-17799da5eeed" (UID: "510e1af6-4069-48d0-a02c-17799da5eeed"). InnerVolumeSpecName "kube-api-access-v69k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.903834 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.971901 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/510e1af6-4069-48d0-a02c-17799da5eeed-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.972125 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/510e1af6-4069-48d0-a02c-17799da5eeed-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:11 crc kubenswrapper[4696]: I1125 11:00:11.972356 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v69k6\" (UniqueName: \"kubernetes.io/projected/510e1af6-4069-48d0-a02c-17799da5eeed-kube-api-access-v69k6\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:12 crc kubenswrapper[4696]: I1125 11:00:12.351124 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" event={"ID":"510e1af6-4069-48d0-a02c-17799da5eeed","Type":"ContainerDied","Data":"3885679d0c3a2c4100021b2e0682a383932239e8e68583039030cbe384b5d3a5"} Nov 25 11:00:12 crc kubenswrapper[4696]: I1125 11:00:12.351160 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3885679d0c3a2c4100021b2e0682a383932239e8e68583039030cbe384b5d3a5" Nov 25 11:00:12 crc kubenswrapper[4696]: I1125 11:00:12.351193 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl" Nov 25 11:00:16 crc kubenswrapper[4696]: I1125 11:00:16.545390 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 11:00:16 crc kubenswrapper[4696]: I1125 11:00:16.596366 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 11:00:17 crc kubenswrapper[4696]: I1125 11:00:17.116459 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dm6hs"] Nov 25 11:00:17 crc kubenswrapper[4696]: I1125 11:00:17.493493 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-svx82" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:17 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:17 crc kubenswrapper[4696]: > Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.403579 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dm6hs" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="registry-server" containerID="cri-o://65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b" gracePeriod=2 Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.692023 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.749689 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.811614 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.866277 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.866330 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.899046 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-utilities\") pod \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.899164 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwc94\" (UniqueName: \"kubernetes.io/projected/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-kube-api-access-jwc94\") pod \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.899193 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-catalog-content\") pod \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\" (UID: \"b17781fb-b2c0-4e7c-8baf-3c43c84c6448\") " Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.899488 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-utilities" (OuterVolumeSpecName: "utilities") pod "b17781fb-b2c0-4e7c-8baf-3c43c84c6448" (UID: "b17781fb-b2c0-4e7c-8baf-3c43c84c6448"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.900389 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.916930 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-kube-api-access-jwc94" (OuterVolumeSpecName: "kube-api-access-jwc94") pod "b17781fb-b2c0-4e7c-8baf-3c43c84c6448" (UID: "b17781fb-b2c0-4e7c-8baf-3c43c84c6448"). InnerVolumeSpecName "kube-api-access-jwc94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:18 crc kubenswrapper[4696]: I1125 11:00:18.950542 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b17781fb-b2c0-4e7c-8baf-3c43c84c6448" (UID: "b17781fb-b2c0-4e7c-8baf-3c43c84c6448"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.002178 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwc94\" (UniqueName: \"kubernetes.io/projected/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-kube-api-access-jwc94\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.002220 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17781fb-b2c0-4e7c-8baf-3c43c84c6448-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.414989 4696 generic.go:334] "Generic (PLEG): container finished" podID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerID="65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b" exitCode=0 Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.416107 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerDied","Data":"65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b"} Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.416203 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dm6hs" event={"ID":"b17781fb-b2c0-4e7c-8baf-3c43c84c6448","Type":"ContainerDied","Data":"d695344da4838e1aae94f483d517e302150964a8e42136c1b673edaa5c54ba87"} Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.416227 4696 scope.go:117] "RemoveContainer" containerID="65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.416186 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dm6hs" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.449643 4696 scope.go:117] "RemoveContainer" containerID="ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.459802 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dm6hs"] Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.468028 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dm6hs"] Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.474170 4696 scope.go:117] "RemoveContainer" containerID="2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.538521 4696 scope.go:117] "RemoveContainer" containerID="65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b" Nov 25 11:00:19 crc kubenswrapper[4696]: E1125 11:00:19.539029 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b\": container with ID starting with 65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b not found: ID does not exist" containerID="65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.539101 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b"} err="failed to get container status \"65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b\": rpc error: code = NotFound desc = could not find container \"65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b\": container with ID starting with 65a21880bd57ca980b6c0880a60b5b1d87a94356981e91c55a4afba12eff9d6b not found: ID does not exist" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.539134 4696 scope.go:117] "RemoveContainer" containerID="ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d" Nov 25 11:00:19 crc kubenswrapper[4696]: E1125 11:00:19.539795 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d\": container with ID starting with ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d not found: ID does not exist" containerID="ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.539832 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d"} err="failed to get container status \"ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d\": rpc error: code = NotFound desc = could not find container \"ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d\": container with ID starting with ed2eb46ad893eaefeacf9e1ec62d0983b1df64151bbe07ef6221ed35e81d673d not found: ID does not exist" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.539851 4696 scope.go:117] "RemoveContainer" containerID="2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de" Nov 25 11:00:19 crc kubenswrapper[4696]: E1125 11:00:19.540211 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de\": container with ID starting with 2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de not found: ID does not exist" containerID="2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.540260 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de"} err="failed to get container status \"2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de\": rpc error: code = NotFound desc = could not find container \"2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de\": container with ID starting with 2fc37150c9272b642090fc8a2224fd317d99fe62aef1ff76a0d4cd9b7f8ee5de not found: ID does not exist" Nov 25 11:00:19 crc kubenswrapper[4696]: I1125 11:00:19.913189 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6xg6h" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:19 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:19 crc kubenswrapper[4696]: > Nov 25 11:00:20 crc kubenswrapper[4696]: I1125 11:00:20.053556 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" path="/var/lib/kubelet/pods/b17781fb-b2c0-4e7c-8baf-3c43c84c6448/volumes" Nov 25 11:00:20 crc kubenswrapper[4696]: I1125 11:00:20.509218 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxf7j"] Nov 25 11:00:20 crc kubenswrapper[4696]: I1125 11:00:20.509424 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qxf7j" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="registry-server" containerID="cri-o://2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32" gracePeriod=2 Nov 25 11:00:20 crc kubenswrapper[4696]: I1125 11:00:20.965255 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.042780 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-catalog-content\") pod \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.042992 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-utilities\") pod \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.043058 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmm8t\" (UniqueName: \"kubernetes.io/projected/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-kube-api-access-pmm8t\") pod \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\" (UID: \"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2\") " Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.043979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-utilities" (OuterVolumeSpecName: "utilities") pod "b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" (UID: "b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.047850 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-kube-api-access-pmm8t" (OuterVolumeSpecName: "kube-api-access-pmm8t") pod "b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" (UID: "b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2"). InnerVolumeSpecName "kube-api-access-pmm8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.066894 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" (UID: "b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.145197 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.145229 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmm8t\" (UniqueName: \"kubernetes.io/projected/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-kube-api-access-pmm8t\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.145240 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.439711 4696 generic.go:334] "Generic (PLEG): container finished" podID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerID="2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32" exitCode=0 Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.439769 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerDied","Data":"2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32"} Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.439816 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxf7j" event={"ID":"b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2","Type":"ContainerDied","Data":"cc834ec7c61acb1e86dc0691f5202ec9f00f659b9a16baf00f5284ddfb5d5f4e"} Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.439849 4696 scope.go:117] "RemoveContainer" containerID="2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.440013 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxf7j" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.497776 4696 scope.go:117] "RemoveContainer" containerID="120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.508288 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxf7j"] Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.524287 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxf7j"] Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.527713 4696 scope.go:117] "RemoveContainer" containerID="de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.573267 4696 scope.go:117] "RemoveContainer" containerID="2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32" Nov 25 11:00:21 crc kubenswrapper[4696]: E1125 11:00:21.573599 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32\": container with ID starting with 2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32 not found: ID does not exist" containerID="2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.573632 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32"} err="failed to get container status \"2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32\": rpc error: code = NotFound desc = could not find container \"2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32\": container with ID starting with 2c5277b329c98faa4bb1e783fb70c6a1030a4fc7cd592b6abbaf077b5e0e1f32 not found: ID does not exist" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.573659 4696 scope.go:117] "RemoveContainer" containerID="120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7" Nov 25 11:00:21 crc kubenswrapper[4696]: E1125 11:00:21.574055 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7\": container with ID starting with 120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7 not found: ID does not exist" containerID="120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.574094 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7"} err="failed to get container status \"120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7\": rpc error: code = NotFound desc = could not find container \"120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7\": container with ID starting with 120965f0eddbfad456b26059c7e42f92798a29c82563f1823d412aec963942c7 not found: ID does not exist" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.574123 4696 scope.go:117] "RemoveContainer" containerID="de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86" Nov 25 11:00:21 crc kubenswrapper[4696]: E1125 11:00:21.574388 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86\": container with ID starting with de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86 not found: ID does not exist" containerID="de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86" Nov 25 11:00:21 crc kubenswrapper[4696]: I1125 11:00:21.574417 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86"} err="failed to get container status \"de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86\": rpc error: code = NotFound desc = could not find container \"de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86\": container with ID starting with de82b70fb3e09785163e5855de6153d551018b9eddb06f3954cc99dfd84b8f86 not found: ID does not exist" Nov 25 11:00:22 crc kubenswrapper[4696]: I1125 11:00:22.057331 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" path="/var/lib/kubelet/pods/b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2/volumes" Nov 25 11:00:23 crc kubenswrapper[4696]: I1125 11:00:23.460677 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" event={"ID":"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a","Type":"ContainerStarted","Data":"ccca62ca43d070ddcca921c7d54c99a763da60bb10de4bcdbb68489dc303ac22"} Nov 25 11:00:26 crc kubenswrapper[4696]: I1125 11:00:26.491065 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svx82" Nov 25 11:00:26 crc kubenswrapper[4696]: I1125 11:00:26.511706 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" podStartSLOduration=7.008180608 podStartE2EDuration="36.511652102s" podCreationTimestamp="2025-11-25 10:59:50 +0000 UTC" firstStartedPulling="2025-11-25 10:59:52.984989475 +0000 UTC m=+1389.798606062" lastFinishedPulling="2025-11-25 11:00:22.488460969 +0000 UTC m=+1419.302077556" observedRunningTime="2025-11-25 11:00:23.478131487 +0000 UTC m=+1420.291748074" watchObservedRunningTime="2025-11-25 11:00:26.511652102 +0000 UTC m=+1423.325268689" Nov 25 11:00:26 crc kubenswrapper[4696]: I1125 11:00:26.544975 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svx82" Nov 25 11:00:26 crc kubenswrapper[4696]: I1125 11:00:26.727763 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svx82"] Nov 25 11:00:28 crc kubenswrapper[4696]: I1125 11:00:28.502408 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svx82" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="registry-server" containerID="cri-o://80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2" gracePeriod=2 Nov 25 11:00:28 crc kubenswrapper[4696]: I1125 11:00:28.991073 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svx82" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.096065 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-catalog-content\") pod \"f7035564-90f9-4dae-ac7a-475b8782825b\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.096327 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csdpr\" (UniqueName: \"kubernetes.io/projected/f7035564-90f9-4dae-ac7a-475b8782825b-kube-api-access-csdpr\") pod \"f7035564-90f9-4dae-ac7a-475b8782825b\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.096378 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-utilities\") pod \"f7035564-90f9-4dae-ac7a-475b8782825b\" (UID: \"f7035564-90f9-4dae-ac7a-475b8782825b\") " Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.096927 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-utilities" (OuterVolumeSpecName: "utilities") pod "f7035564-90f9-4dae-ac7a-475b8782825b" (UID: "f7035564-90f9-4dae-ac7a-475b8782825b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.099391 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.101942 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7035564-90f9-4dae-ac7a-475b8782825b-kube-api-access-csdpr" (OuterVolumeSpecName: "kube-api-access-csdpr") pod "f7035564-90f9-4dae-ac7a-475b8782825b" (UID: "f7035564-90f9-4dae-ac7a-475b8782825b"). InnerVolumeSpecName "kube-api-access-csdpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.157481 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7035564-90f9-4dae-ac7a-475b8782825b" (UID: "f7035564-90f9-4dae-ac7a-475b8782825b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.201551 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csdpr\" (UniqueName: \"kubernetes.io/projected/f7035564-90f9-4dae-ac7a-475b8782825b-kube-api-access-csdpr\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.201588 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7035564-90f9-4dae-ac7a-475b8782825b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.517468 4696 generic.go:334] "Generic (PLEG): container finished" podID="f7035564-90f9-4dae-ac7a-475b8782825b" containerID="80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2" exitCode=0 Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.517757 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerDied","Data":"80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2"} Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.517781 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svx82" event={"ID":"f7035564-90f9-4dae-ac7a-475b8782825b","Type":"ContainerDied","Data":"3680ef82a867784082d3199b8d92522525f904a7c5ff68d3d3c7509ad3c4b18b"} Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.517797 4696 scope.go:117] "RemoveContainer" containerID="80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.517954 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svx82" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.555942 4696 scope.go:117] "RemoveContainer" containerID="1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.567981 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svx82"] Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.586319 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svx82"] Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.595035 4696 scope.go:117] "RemoveContainer" containerID="3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.627205 4696 scope.go:117] "RemoveContainer" containerID="80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2" Nov 25 11:00:29 crc kubenswrapper[4696]: E1125 11:00:29.627682 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2\": container with ID starting with 80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2 not found: ID does not exist" containerID="80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.627735 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2"} err="failed to get container status \"80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2\": rpc error: code = NotFound desc = could not find container \"80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2\": container with ID starting with 80f6e9e3ba8ec48657747e8af0aba35da8c82f324d1bc62c958612a897edeef2 not found: ID does not exist" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.627768 4696 scope.go:117] "RemoveContainer" containerID="1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4" Nov 25 11:00:29 crc kubenswrapper[4696]: E1125 11:00:29.628177 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4\": container with ID starting with 1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4 not found: ID does not exist" containerID="1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.628236 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4"} err="failed to get container status \"1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4\": rpc error: code = NotFound desc = could not find container \"1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4\": container with ID starting with 1dd09b6def1c530fec5284ae610e18bd5103b6e869d1a425a97beb2c4ad2d3d4 not found: ID does not exist" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.628275 4696 scope.go:117] "RemoveContainer" containerID="3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938" Nov 25 11:00:29 crc kubenswrapper[4696]: E1125 11:00:29.628892 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938\": container with ID starting with 3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938 not found: ID does not exist" containerID="3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.628924 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938"} err="failed to get container status \"3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938\": rpc error: code = NotFound desc = could not find container \"3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938\": container with ID starting with 3434b1fd8735da7be843bc862fec76a00fd82f8a91683dfbbac29a80a2918938 not found: ID does not exist" Nov 25 11:00:29 crc kubenswrapper[4696]: I1125 11:00:29.920774 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6xg6h" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:29 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:29 crc kubenswrapper[4696]: > Nov 25 11:00:30 crc kubenswrapper[4696]: I1125 11:00:30.052549 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" path="/var/lib/kubelet/pods/f7035564-90f9-4dae-ac7a-475b8782825b/volumes" Nov 25 11:00:30 crc kubenswrapper[4696]: I1125 11:00:30.802234 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:00:30 crc kubenswrapper[4696]: I1125 11:00:30.802316 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:00:33 crc kubenswrapper[4696]: I1125 11:00:33.825554 4696 scope.go:117] "RemoveContainer" containerID="cdb45bee6c0bc35518f191752999d1266038fe12e4b900064b46879e341b322d" Nov 25 11:00:33 crc kubenswrapper[4696]: I1125 11:00:33.861176 4696 scope.go:117] "RemoveContainer" containerID="0f4b495f33536cabfe56fbf85bd87cd45f4b9603bbffaeaca651bbdcbeb8d9e8" Nov 25 11:00:33 crc kubenswrapper[4696]: I1125 11:00:33.895234 4696 scope.go:117] "RemoveContainer" containerID="8f50f81db271f4f47fc5b1a6901dce23cc6399e8c12abe5901b52750df59aa40" Nov 25 11:00:39 crc kubenswrapper[4696]: I1125 11:00:39.611769 4696 generic.go:334] "Generic (PLEG): container finished" podID="4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" containerID="ccca62ca43d070ddcca921c7d54c99a763da60bb10de4bcdbb68489dc303ac22" exitCode=0 Nov 25 11:00:39 crc kubenswrapper[4696]: I1125 11:00:39.611862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" event={"ID":"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a","Type":"ContainerDied","Data":"ccca62ca43d070ddcca921c7d54c99a763da60bb10de4bcdbb68489dc303ac22"} Nov 25 11:00:39 crc kubenswrapper[4696]: I1125 11:00:39.913355 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6xg6h" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:00:39 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:00:39 crc kubenswrapper[4696]: > Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.007126 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.114070 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-ssh-key\") pod \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.114149 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-repo-setup-combined-ca-bundle\") pod \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.114248 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmf6p\" (UniqueName: \"kubernetes.io/projected/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-kube-api-access-cmf6p\") pod \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.114291 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-inventory\") pod \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\" (UID: \"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a\") " Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.121655 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" (UID: "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.121720 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-kube-api-access-cmf6p" (OuterVolumeSpecName: "kube-api-access-cmf6p") pod "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" (UID: "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a"). InnerVolumeSpecName "kube-api-access-cmf6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.145370 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" (UID: "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.150975 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-inventory" (OuterVolumeSpecName: "inventory") pod "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" (UID: "4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.217192 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmf6p\" (UniqueName: \"kubernetes.io/projected/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-kube-api-access-cmf6p\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.217416 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.217425 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.217435 4696 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.630435 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" event={"ID":"4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a","Type":"ContainerDied","Data":"6964730550e50fad669b0263b0291ae5bab5119710a5f01103a7ba4457b4cea5"} Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.630482 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6964730550e50fad669b0263b0291ae5bab5119710a5f01103a7ba4457b4cea5" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.630928 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.733657 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj"] Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734090 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734110 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734125 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="extract-utilities" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734133 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="extract-utilities" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734146 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="extract-utilities" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734154 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="extract-utilities" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734162 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="extract-utilities" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734170 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="extract-utilities" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734186 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734193 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734205 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="extract-content" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734213 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="extract-content" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734228 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="extract-content" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734235 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="extract-content" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734245 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="510e1af6-4069-48d0-a02c-17799da5eeed" containerName="collect-profiles" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734251 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="510e1af6-4069-48d0-a02c-17799da5eeed" containerName="collect-profiles" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734262 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734268 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734279 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="extract-content" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734284 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="extract-content" Nov 25 11:00:41 crc kubenswrapper[4696]: E1125 11:00:41.734295 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734303 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734464 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0eb8e32-cd66-4be0-8bf2-659d4c2ea8c2" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734479 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734494 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7035564-90f9-4dae-ac7a-475b8782825b" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734514 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="510e1af6-4069-48d0-a02c-17799da5eeed" containerName="collect-profiles" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.734532 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17781fb-b2c0-4e7c-8baf-3c43c84c6448" containerName="registry-server" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.735168 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.737728 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.738230 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.738707 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.756561 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.818322 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj"] Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.829286 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g897m\" (UniqueName: \"kubernetes.io/projected/2a97a4bb-2255-4f00-901a-3392d7808c53-kube-api-access-g897m\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.829472 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.829676 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.930977 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.931295 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g897m\" (UniqueName: \"kubernetes.io/projected/2a97a4bb-2255-4f00-901a-3392d7808c53-kube-api-access-g897m\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.931412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.935792 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.935953 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:41 crc kubenswrapper[4696]: I1125 11:00:41.948042 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g897m\" (UniqueName: \"kubernetes.io/projected/2a97a4bb-2255-4f00-901a-3392d7808c53-kube-api-access-g897m\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ls6nj\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:42 crc kubenswrapper[4696]: I1125 11:00:42.061574 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:42 crc kubenswrapper[4696]: I1125 11:00:42.570059 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj"] Nov 25 11:00:42 crc kubenswrapper[4696]: I1125 11:00:42.640551 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" event={"ID":"2a97a4bb-2255-4f00-901a-3392d7808c53","Type":"ContainerStarted","Data":"85a7e30b56e077872303e8ee8ed34383daaacb38a900f87268eb58db4cf08921"} Nov 25 11:00:43 crc kubenswrapper[4696]: I1125 11:00:43.651583 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" event={"ID":"2a97a4bb-2255-4f00-901a-3392d7808c53","Type":"ContainerStarted","Data":"71e0dfd5aec42206229434d1661800ae09bed6d8cff5c0077e49c7dada50fe3a"} Nov 25 11:00:43 crc kubenswrapper[4696]: I1125 11:00:43.675735 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" podStartSLOduration=2.208596478 podStartE2EDuration="2.675717719s" podCreationTimestamp="2025-11-25 11:00:41 +0000 UTC" firstStartedPulling="2025-11-25 11:00:42.579620709 +0000 UTC m=+1439.393237296" lastFinishedPulling="2025-11-25 11:00:43.04674195 +0000 UTC m=+1439.860358537" observedRunningTime="2025-11-25 11:00:43.669756774 +0000 UTC m=+1440.483373371" watchObservedRunningTime="2025-11-25 11:00:43.675717719 +0000 UTC m=+1440.489334326" Nov 25 11:00:46 crc kubenswrapper[4696]: I1125 11:00:46.682694 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" event={"ID":"2a97a4bb-2255-4f00-901a-3392d7808c53","Type":"ContainerDied","Data":"71e0dfd5aec42206229434d1661800ae09bed6d8cff5c0077e49c7dada50fe3a"} Nov 25 11:00:46 crc kubenswrapper[4696]: I1125 11:00:46.682643 4696 generic.go:334] "Generic (PLEG): container finished" podID="2a97a4bb-2255-4f00-901a-3392d7808c53" containerID="71e0dfd5aec42206229434d1661800ae09bed6d8cff5c0077e49c7dada50fe3a" exitCode=0 Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.065997 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.247294 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g897m\" (UniqueName: \"kubernetes.io/projected/2a97a4bb-2255-4f00-901a-3392d7808c53-kube-api-access-g897m\") pod \"2a97a4bb-2255-4f00-901a-3392d7808c53\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.247379 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-inventory\") pod \"2a97a4bb-2255-4f00-901a-3392d7808c53\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.247417 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-ssh-key\") pod \"2a97a4bb-2255-4f00-901a-3392d7808c53\" (UID: \"2a97a4bb-2255-4f00-901a-3392d7808c53\") " Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.252168 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a97a4bb-2255-4f00-901a-3392d7808c53-kube-api-access-g897m" (OuterVolumeSpecName: "kube-api-access-g897m") pod "2a97a4bb-2255-4f00-901a-3392d7808c53" (UID: "2a97a4bb-2255-4f00-901a-3392d7808c53"). InnerVolumeSpecName "kube-api-access-g897m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.274787 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a97a4bb-2255-4f00-901a-3392d7808c53" (UID: "2a97a4bb-2255-4f00-901a-3392d7808c53"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.278853 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-inventory" (OuterVolumeSpecName: "inventory") pod "2a97a4bb-2255-4f00-901a-3392d7808c53" (UID: "2a97a4bb-2255-4f00-901a-3392d7808c53"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.349259 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g897m\" (UniqueName: \"kubernetes.io/projected/2a97a4bb-2255-4f00-901a-3392d7808c53-kube-api-access-g897m\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.349297 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.349306 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a97a4bb-2255-4f00-901a-3392d7808c53-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.704206 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" event={"ID":"2a97a4bb-2255-4f00-901a-3392d7808c53","Type":"ContainerDied","Data":"85a7e30b56e077872303e8ee8ed34383daaacb38a900f87268eb58db4cf08921"} Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.704410 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85a7e30b56e077872303e8ee8ed34383daaacb38a900f87268eb58db4cf08921" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.704294 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ls6nj" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.795087 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952"] Nov 25 11:00:48 crc kubenswrapper[4696]: E1125 11:00:48.795478 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a97a4bb-2255-4f00-901a-3392d7808c53" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.795494 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a97a4bb-2255-4f00-901a-3392d7808c53" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.795658 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a97a4bb-2255-4f00-901a-3392d7808c53" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.796708 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.799101 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.800172 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.800430 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.805988 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.812657 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952"] Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.859060 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.859352 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.859559 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2szg\" (UniqueName: \"kubernetes.io/projected/062b8e49-6137-4b89-8ae4-a29272ff0d1f-kube-api-access-z2szg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.859633 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.911036 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.958573 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.961947 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.961996 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.962069 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2szg\" (UniqueName: \"kubernetes.io/projected/062b8e49-6137-4b89-8ae4-a29272ff0d1f-kube-api-access-z2szg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.962109 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.967097 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.967111 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.967262 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:48 crc kubenswrapper[4696]: I1125 11:00:48.981288 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2szg\" (UniqueName: \"kubernetes.io/projected/062b8e49-6137-4b89-8ae4-a29272ff0d1f-kube-api-access-z2szg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-65952\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:49 crc kubenswrapper[4696]: I1125 11:00:49.116028 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:00:49 crc kubenswrapper[4696]: I1125 11:00:49.171558 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xg6h"] Nov 25 11:00:49 crc kubenswrapper[4696]: I1125 11:00:49.456552 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952"] Nov 25 11:00:49 crc kubenswrapper[4696]: I1125 11:00:49.715027 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" event={"ID":"062b8e49-6137-4b89-8ae4-a29272ff0d1f","Type":"ContainerStarted","Data":"a06bd65c33cabce312aad4230c6e2de2e4ae50f4c7c3d9b38ff7ffbfdb38a49a"} Nov 25 11:00:50 crc kubenswrapper[4696]: I1125 11:00:50.727391 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" event={"ID":"062b8e49-6137-4b89-8ae4-a29272ff0d1f","Type":"ContainerStarted","Data":"d3c56f899bf59df3ef8dc0077afdf6a9bccf089a6a34eec458a90077863691d5"} Nov 25 11:00:50 crc kubenswrapper[4696]: I1125 11:00:50.727525 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6xg6h" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" containerID="cri-o://065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4" gracePeriod=2 Nov 25 11:00:50 crc kubenswrapper[4696]: I1125 11:00:50.784218 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" podStartSLOduration=2.286172128 podStartE2EDuration="2.78418831s" podCreationTimestamp="2025-11-25 11:00:48 +0000 UTC" firstStartedPulling="2025-11-25 11:00:49.471295351 +0000 UTC m=+1446.284911938" lastFinishedPulling="2025-11-25 11:00:49.969311533 +0000 UTC m=+1446.782928120" observedRunningTime="2025-11-25 11:00:50.768831652 +0000 UTC m=+1447.582448239" watchObservedRunningTime="2025-11-25 11:00:50.78418831 +0000 UTC m=+1447.597804917" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.219410 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.304941 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjkfs\" (UniqueName: \"kubernetes.io/projected/2fd52a93-18b2-424b-8a35-d09d44f47d7b-kube-api-access-bjkfs\") pod \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.304982 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-catalog-content\") pod \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.305107 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-utilities\") pod \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\" (UID: \"2fd52a93-18b2-424b-8a35-d09d44f47d7b\") " Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.305811 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-utilities" (OuterVolumeSpecName: "utilities") pod "2fd52a93-18b2-424b-8a35-d09d44f47d7b" (UID: "2fd52a93-18b2-424b-8a35-d09d44f47d7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.313658 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd52a93-18b2-424b-8a35-d09d44f47d7b-kube-api-access-bjkfs" (OuterVolumeSpecName: "kube-api-access-bjkfs") pod "2fd52a93-18b2-424b-8a35-d09d44f47d7b" (UID: "2fd52a93-18b2-424b-8a35-d09d44f47d7b"). InnerVolumeSpecName "kube-api-access-bjkfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.407051 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fd52a93-18b2-424b-8a35-d09d44f47d7b" (UID: "2fd52a93-18b2-424b-8a35-d09d44f47d7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.408455 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjkfs\" (UniqueName: \"kubernetes.io/projected/2fd52a93-18b2-424b-8a35-d09d44f47d7b-kube-api-access-bjkfs\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.408493 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.408502 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd52a93-18b2-424b-8a35-d09d44f47d7b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.740278 4696 generic.go:334] "Generic (PLEG): container finished" podID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerID="065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4" exitCode=0 Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.740346 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xg6h" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.740390 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerDied","Data":"065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4"} Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.740440 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xg6h" event={"ID":"2fd52a93-18b2-424b-8a35-d09d44f47d7b","Type":"ContainerDied","Data":"fea8f5ba69957dc742dd47db8bd4156b1a229567144fda98716669422105bbb0"} Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.740459 4696 scope.go:117] "RemoveContainer" containerID="065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.765001 4696 scope.go:117] "RemoveContainer" containerID="da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.776121 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xg6h"] Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.786511 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6xg6h"] Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.805164 4696 scope.go:117] "RemoveContainer" containerID="6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.832899 4696 scope.go:117] "RemoveContainer" containerID="065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4" Nov 25 11:00:51 crc kubenswrapper[4696]: E1125 11:00:51.833347 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4\": container with ID starting with 065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4 not found: ID does not exist" containerID="065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.833387 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4"} err="failed to get container status \"065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4\": rpc error: code = NotFound desc = could not find container \"065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4\": container with ID starting with 065c67ab7ab9998addac0210d3617206b3e83391b3f60f85f79452b9de3cd2d4 not found: ID does not exist" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.833418 4696 scope.go:117] "RemoveContainer" containerID="da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b" Nov 25 11:00:51 crc kubenswrapper[4696]: E1125 11:00:51.833776 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b\": container with ID starting with da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b not found: ID does not exist" containerID="da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.833837 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b"} err="failed to get container status \"da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b\": rpc error: code = NotFound desc = could not find container \"da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b\": container with ID starting with da8c62e3519c6a7c39190d5c4bde0a89556f16ff7d8901975bf2ab8b1db0617b not found: ID does not exist" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.833857 4696 scope.go:117] "RemoveContainer" containerID="6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1" Nov 25 11:00:51 crc kubenswrapper[4696]: E1125 11:00:51.834253 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1\": container with ID starting with 6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1 not found: ID does not exist" containerID="6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1" Nov 25 11:00:51 crc kubenswrapper[4696]: I1125 11:00:51.834283 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1"} err="failed to get container status \"6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1\": rpc error: code = NotFound desc = could not find container \"6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1\": container with ID starting with 6f0e583f0a1ad879f848d301feba929e61e17895f4494256eedb8f7b074e54d1 not found: ID does not exist" Nov 25 11:00:52 crc kubenswrapper[4696]: I1125 11:00:52.059619 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" path="/var/lib/kubelet/pods/2fd52a93-18b2-424b-8a35-d09d44f47d7b/volumes" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.138197 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401141-q8gxc"] Nov 25 11:01:00 crc kubenswrapper[4696]: E1125 11:01:00.139178 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="extract-utilities" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.139195 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="extract-utilities" Nov 25 11:01:00 crc kubenswrapper[4696]: E1125 11:01:00.139225 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.139233 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" Nov 25 11:01:00 crc kubenswrapper[4696]: E1125 11:01:00.139254 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="extract-content" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.139262 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="extract-content" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.139506 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd52a93-18b2-424b-8a35-d09d44f47d7b" containerName="registry-server" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.140499 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.172711 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401141-q8gxc"] Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.272494 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-config-data\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.272845 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c27d4\" (UniqueName: \"kubernetes.io/projected/0a3da9a0-6262-4ff8-b510-e476db9d9985-kube-api-access-c27d4\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.272865 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-combined-ca-bundle\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.272928 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-fernet-keys\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.374452 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-config-data\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.374507 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c27d4\" (UniqueName: \"kubernetes.io/projected/0a3da9a0-6262-4ff8-b510-e476db9d9985-kube-api-access-c27d4\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.374545 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-combined-ca-bundle\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.374631 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-fernet-keys\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.381026 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-config-data\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.381211 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-combined-ca-bundle\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.383441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-fernet-keys\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.393681 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c27d4\" (UniqueName: \"kubernetes.io/projected/0a3da9a0-6262-4ff8-b510-e476db9d9985-kube-api-access-c27d4\") pod \"keystone-cron-29401141-q8gxc\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.469906 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.801656 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.802023 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:01:00 crc kubenswrapper[4696]: I1125 11:01:00.915704 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401141-q8gxc"] Nov 25 11:01:01 crc kubenswrapper[4696]: I1125 11:01:01.868548 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-q8gxc" event={"ID":"0a3da9a0-6262-4ff8-b510-e476db9d9985","Type":"ContainerStarted","Data":"52fb5324886149a948ab7fe94fc5c1bbac414750c523c4d73268aad59d4d7ab4"} Nov 25 11:01:01 crc kubenswrapper[4696]: I1125 11:01:01.869780 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-q8gxc" event={"ID":"0a3da9a0-6262-4ff8-b510-e476db9d9985","Type":"ContainerStarted","Data":"1ee38ab94289058b7155a07c5026e017bad19972e671aa97f1743fed792365e3"} Nov 25 11:01:01 crc kubenswrapper[4696]: I1125 11:01:01.892807 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401141-q8gxc" podStartSLOduration=1.89278089 podStartE2EDuration="1.89278089s" podCreationTimestamp="2025-11-25 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:01.885539575 +0000 UTC m=+1458.699156162" watchObservedRunningTime="2025-11-25 11:01:01.89278089 +0000 UTC m=+1458.706397497" Nov 25 11:01:05 crc kubenswrapper[4696]: I1125 11:01:05.906326 4696 generic.go:334] "Generic (PLEG): container finished" podID="0a3da9a0-6262-4ff8-b510-e476db9d9985" containerID="52fb5324886149a948ab7fe94fc5c1bbac414750c523c4d73268aad59d4d7ab4" exitCode=0 Nov 25 11:01:05 crc kubenswrapper[4696]: I1125 11:01:05.906415 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-q8gxc" event={"ID":"0a3da9a0-6262-4ff8-b510-e476db9d9985","Type":"ContainerDied","Data":"52fb5324886149a948ab7fe94fc5c1bbac414750c523c4d73268aad59d4d7ab4"} Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.208357 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.313636 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c27d4\" (UniqueName: \"kubernetes.io/projected/0a3da9a0-6262-4ff8-b510-e476db9d9985-kube-api-access-c27d4\") pod \"0a3da9a0-6262-4ff8-b510-e476db9d9985\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.313769 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-combined-ca-bundle\") pod \"0a3da9a0-6262-4ff8-b510-e476db9d9985\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.313862 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-fernet-keys\") pod \"0a3da9a0-6262-4ff8-b510-e476db9d9985\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.313917 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-config-data\") pod \"0a3da9a0-6262-4ff8-b510-e476db9d9985\" (UID: \"0a3da9a0-6262-4ff8-b510-e476db9d9985\") " Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.319497 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0a3da9a0-6262-4ff8-b510-e476db9d9985" (UID: "0a3da9a0-6262-4ff8-b510-e476db9d9985"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.319725 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a3da9a0-6262-4ff8-b510-e476db9d9985-kube-api-access-c27d4" (OuterVolumeSpecName: "kube-api-access-c27d4") pod "0a3da9a0-6262-4ff8-b510-e476db9d9985" (UID: "0a3da9a0-6262-4ff8-b510-e476db9d9985"). InnerVolumeSpecName "kube-api-access-c27d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.348392 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a3da9a0-6262-4ff8-b510-e476db9d9985" (UID: "0a3da9a0-6262-4ff8-b510-e476db9d9985"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.369299 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-config-data" (OuterVolumeSpecName: "config-data") pod "0a3da9a0-6262-4ff8-b510-e476db9d9985" (UID: "0a3da9a0-6262-4ff8-b510-e476db9d9985"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.415832 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c27d4\" (UniqueName: \"kubernetes.io/projected/0a3da9a0-6262-4ff8-b510-e476db9d9985-kube-api-access-c27d4\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.415868 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.415880 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.415891 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a3da9a0-6262-4ff8-b510-e476db9d9985-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.927655 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-q8gxc" event={"ID":"0a3da9a0-6262-4ff8-b510-e476db9d9985","Type":"ContainerDied","Data":"1ee38ab94289058b7155a07c5026e017bad19972e671aa97f1743fed792365e3"} Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.927981 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee38ab94289058b7155a07c5026e017bad19972e671aa97f1743fed792365e3" Nov 25 11:01:07 crc kubenswrapper[4696]: I1125 11:01:07.927735 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-q8gxc" Nov 25 11:01:30 crc kubenswrapper[4696]: I1125 11:01:30.802478 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:01:30 crc kubenswrapper[4696]: I1125 11:01:30.803491 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:01:30 crc kubenswrapper[4696]: I1125 11:01:30.803587 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:01:30 crc kubenswrapper[4696]: I1125 11:01:30.804766 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a6788bdae639f3c35f45af42113cd2f9e953a8da542ed0cc824a9b88a8dd56b"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:01:30 crc kubenswrapper[4696]: I1125 11:01:30.804835 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://8a6788bdae639f3c35f45af42113cd2f9e953a8da542ed0cc824a9b88a8dd56b" gracePeriod=600 Nov 25 11:01:31 crc kubenswrapper[4696]: I1125 11:01:31.165402 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="8a6788bdae639f3c35f45af42113cd2f9e953a8da542ed0cc824a9b88a8dd56b" exitCode=0 Nov 25 11:01:31 crc kubenswrapper[4696]: I1125 11:01:31.165451 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"8a6788bdae639f3c35f45af42113cd2f9e953a8da542ed0cc824a9b88a8dd56b"} Nov 25 11:01:31 crc kubenswrapper[4696]: I1125 11:01:31.165808 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9"} Nov 25 11:01:31 crc kubenswrapper[4696]: I1125 11:01:31.165833 4696 scope.go:117] "RemoveContainer" containerID="b5fa27a9f95f70b987f66a1cc4d9ba856c14c78b0befd578a595d382c8833769" Nov 25 11:01:34 crc kubenswrapper[4696]: I1125 11:01:34.086018 4696 scope.go:117] "RemoveContainer" containerID="c9bedf8f9d337dffdbf2d4e0c891453886448ec59a81619beb91ae723ecca0dc" Nov 25 11:03:34 crc kubenswrapper[4696]: I1125 11:03:34.222229 4696 scope.go:117] "RemoveContainer" containerID="30810c9e3eecbf9e1be83ee7ada1ac9b5ccadf967e125fb72e9068e4d53c64f7" Nov 25 11:03:34 crc kubenswrapper[4696]: I1125 11:03:34.265033 4696 scope.go:117] "RemoveContainer" containerID="bceb37e2683d4e9ee659c7172a432555d7927e553779cdaced918cb5dbad1b2f" Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.064869 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-924c-account-create-t4bcj"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.085253 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5ebe-account-create-7ffls"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.096507 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-p47td"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.105491 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-d9gr4"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.114307 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-63be-account-create-jxwx5"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.123205 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-d9gr4"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.133612 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-924c-account-create-t4bcj"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.141633 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-p47td"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.149587 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5ebe-account-create-7ffls"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.158307 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-63be-account-create-jxwx5"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.167108 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-xb2zp"] Nov 25 11:03:53 crc kubenswrapper[4696]: I1125 11:03:53.174936 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-xb2zp"] Nov 25 11:03:54 crc kubenswrapper[4696]: I1125 11:03:54.055237 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00abba76-7f70-4cb6-a4fa-46a44e2ebaac" path="/var/lib/kubelet/pods/00abba76-7f70-4cb6-a4fa-46a44e2ebaac/volumes" Nov 25 11:03:54 crc kubenswrapper[4696]: I1125 11:03:54.058502 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e192cbe-f2c6-44c0-ba41-81bf3e14705d" path="/var/lib/kubelet/pods/1e192cbe-f2c6-44c0-ba41-81bf3e14705d/volumes" Nov 25 11:03:54 crc kubenswrapper[4696]: I1125 11:03:54.060681 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2815f305-2a4d-456e-af2c-0c842eda6acf" path="/var/lib/kubelet/pods/2815f305-2a4d-456e-af2c-0c842eda6acf/volumes" Nov 25 11:03:54 crc kubenswrapper[4696]: I1125 11:03:54.065064 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7631c79c-d022-41b6-916e-102fcc294874" path="/var/lib/kubelet/pods/7631c79c-d022-41b6-916e-102fcc294874/volumes" Nov 25 11:03:54 crc kubenswrapper[4696]: I1125 11:03:54.067071 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ff86ec-50a8-4137-8f0e-b1adb75ffe34" path="/var/lib/kubelet/pods/c5ff86ec-50a8-4137-8f0e-b1adb75ffe34/volumes" Nov 25 11:03:54 crc kubenswrapper[4696]: I1125 11:03:54.068947 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca998c2e-0f79-4bbe-8095-897dc135007c" path="/var/lib/kubelet/pods/ca998c2e-0f79-4bbe-8095-897dc135007c/volumes" Nov 25 11:04:00 crc kubenswrapper[4696]: I1125 11:04:00.801736 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:04:00 crc kubenswrapper[4696]: I1125 11:04:00.802798 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.055561 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-401d-account-create-ptvsr"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.066713 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d0d4-account-create-qhm9b"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.077655 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-bp7v2"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.091824 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b1f9-account-create-4pwnx"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.100124 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qdf6p"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.107681 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xtjbt"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.115419 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b1f9-account-create-4pwnx"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.123845 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-401d-account-create-ptvsr"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.131864 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xtjbt"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.140920 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d0d4-account-create-qhm9b"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.153111 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-bp7v2"] Nov 25 11:04:26 crc kubenswrapper[4696]: I1125 11:04:26.161460 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qdf6p"] Nov 25 11:04:28 crc kubenswrapper[4696]: I1125 11:04:28.053590 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c214c5b-9ca0-4499-814e-65a5fa45e42a" path="/var/lib/kubelet/pods/1c214c5b-9ca0-4499-814e-65a5fa45e42a/volumes" Nov 25 11:04:28 crc kubenswrapper[4696]: I1125 11:04:28.063636 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61fbc9f2-4e89-4d11-9d92-434700854f9f" path="/var/lib/kubelet/pods/61fbc9f2-4e89-4d11-9d92-434700854f9f/volumes" Nov 25 11:04:28 crc kubenswrapper[4696]: I1125 11:04:28.071736 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68e586e5-acb4-4bc7-be27-545032778925" path="/var/lib/kubelet/pods/68e586e5-acb4-4bc7-be27-545032778925/volumes" Nov 25 11:04:28 crc kubenswrapper[4696]: I1125 11:04:28.073504 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5483447-69b5-485d-8ea2-2de150da78bb" path="/var/lib/kubelet/pods/b5483447-69b5-485d-8ea2-2de150da78bb/volumes" Nov 25 11:04:28 crc kubenswrapper[4696]: I1125 11:04:28.075029 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e389cdae-22d0-4285-8f08-d26bda099503" path="/var/lib/kubelet/pods/e389cdae-22d0-4285-8f08-d26bda099503/volumes" Nov 25 11:04:28 crc kubenswrapper[4696]: I1125 11:04:28.076524 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa48aa3d-d214-4690-8a29-a140a2c11d77" path="/var/lib/kubelet/pods/fa48aa3d-d214-4690-8a29-a140a2c11d77/volumes" Nov 25 11:04:30 crc kubenswrapper[4696]: I1125 11:04:30.801824 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:04:30 crc kubenswrapper[4696]: I1125 11:04:30.802137 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:04:32 crc kubenswrapper[4696]: I1125 11:04:32.028309 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4xkzg"] Nov 25 11:04:32 crc kubenswrapper[4696]: I1125 11:04:32.037229 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4xkzg"] Nov 25 11:04:32 crc kubenswrapper[4696]: I1125 11:04:32.076792 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143d89a1-ef2c-4e93-8bd7-40e8a6044bca" path="/var/lib/kubelet/pods/143d89a1-ef2c-4e93-8bd7-40e8a6044bca/volumes" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.329541 4696 scope.go:117] "RemoveContainer" containerID="a1e21147c6240a548d496dccb577f9e853613f365e7d7265b9d79e18445e106e" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.353718 4696 scope.go:117] "RemoveContainer" containerID="257f611c064b71ea589c01025446c79f05ee6e3c5c6279ce562e5b2efbca806d" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.383996 4696 scope.go:117] "RemoveContainer" containerID="1d45ffde46f1f111f12acbf0772017a3612d4ca37705c85102446cf1a0c8faf8" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.457853 4696 scope.go:117] "RemoveContainer" containerID="1f446d8e3bd63349999c9b05cdcbc15544c2da06e464b752e13b25ef43e49bf8" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.487695 4696 scope.go:117] "RemoveContainer" containerID="0a3edf2ba429339228ed5782c0e0039a92d3dfaf39806cc4e9f1515032854441" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.542517 4696 scope.go:117] "RemoveContainer" containerID="f0d69789e1b7a66149e709d8a9c9e9c5497bb3d8ca3dbe145dacdde09fb24660" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.603540 4696 scope.go:117] "RemoveContainer" containerID="d61549e48123092e2468424e4ab44901eb7c82dda018ed09824954bc88803735" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.651384 4696 scope.go:117] "RemoveContainer" containerID="574ac7dfca7b3f18af5d2c4c4886c1374979b15a368d42c096248a8c7c1607da" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.674541 4696 scope.go:117] "RemoveContainer" containerID="66a23011e15065483610e5debb6babcdae1dfb3c4d1bc930ca2d74b7dd80fa29" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.700759 4696 scope.go:117] "RemoveContainer" containerID="318d4f95c617c7edcb581ef02df1cde2006040322c6877f5f5babd4e33a4ef63" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.726767 4696 scope.go:117] "RemoveContainer" containerID="1cdef9c1ffa9760a5e1318ecc1d79b3504841ca8e2ca6f7382c6f5fedf2011e5" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.746542 4696 scope.go:117] "RemoveContainer" containerID="fddf84327a2616525b0c6a26d45acab7c9e3acb6b06562585e089f24a871fff9" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.772262 4696 scope.go:117] "RemoveContainer" containerID="4c6005b993b3e6be023724f76800a565e233a0d13ad0c770ac718e2d07143625" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.807443 4696 scope.go:117] "RemoveContainer" containerID="e8332f98c4246923175ad3a8fc28967629dbe8d7145c7c23d77cb8507acca34b" Nov 25 11:04:34 crc kubenswrapper[4696]: I1125 11:04:34.836847 4696 scope.go:117] "RemoveContainer" containerID="a8947ee445d01d5469d2336c922f90c9c89ec1d50ff83b761cbf2a20cabcfef4" Nov 25 11:04:56 crc kubenswrapper[4696]: I1125 11:04:56.052285 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ljbtx"] Nov 25 11:04:56 crc kubenswrapper[4696]: I1125 11:04:56.060400 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ljbtx"] Nov 25 11:04:58 crc kubenswrapper[4696]: I1125 11:04:58.054763 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2593ea-7985-4050-8000-2be38f4b488c" path="/var/lib/kubelet/pods/bc2593ea-7985-4050-8000-2be38f4b488c/volumes" Nov 25 11:05:00 crc kubenswrapper[4696]: I1125 11:05:00.802118 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:05:00 crc kubenswrapper[4696]: I1125 11:05:00.802447 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:05:00 crc kubenswrapper[4696]: I1125 11:05:00.802516 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:05:00 crc kubenswrapper[4696]: I1125 11:05:00.803186 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:05:00 crc kubenswrapper[4696]: I1125 11:05:00.803243 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" gracePeriod=600 Nov 25 11:05:01 crc kubenswrapper[4696]: I1125 11:05:01.068921 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" exitCode=0 Nov 25 11:05:01 crc kubenswrapper[4696]: I1125 11:05:01.068960 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9"} Nov 25 11:05:01 crc kubenswrapper[4696]: I1125 11:05:01.069309 4696 scope.go:117] "RemoveContainer" containerID="8a6788bdae639f3c35f45af42113cd2f9e953a8da542ed0cc824a9b88a8dd56b" Nov 25 11:05:01 crc kubenswrapper[4696]: E1125 11:05:01.504239 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:05:02 crc kubenswrapper[4696]: I1125 11:05:02.080768 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:05:02 crc kubenswrapper[4696]: E1125 11:05:02.081407 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:05:05 crc kubenswrapper[4696]: I1125 11:05:05.114539 4696 generic.go:334] "Generic (PLEG): container finished" podID="062b8e49-6137-4b89-8ae4-a29272ff0d1f" containerID="d3c56f899bf59df3ef8dc0077afdf6a9bccf089a6a34eec458a90077863691d5" exitCode=0 Nov 25 11:05:05 crc kubenswrapper[4696]: I1125 11:05:05.114647 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" event={"ID":"062b8e49-6137-4b89-8ae4-a29272ff0d1f","Type":"ContainerDied","Data":"d3c56f899bf59df3ef8dc0077afdf6a9bccf089a6a34eec458a90077863691d5"} Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.560779 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.611138 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-inventory\") pod \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.611411 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-bootstrap-combined-ca-bundle\") pod \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.611499 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-ssh-key\") pod \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.611707 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2szg\" (UniqueName: \"kubernetes.io/projected/062b8e49-6137-4b89-8ae4-a29272ff0d1f-kube-api-access-z2szg\") pod \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\" (UID: \"062b8e49-6137-4b89-8ae4-a29272ff0d1f\") " Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.617587 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "062b8e49-6137-4b89-8ae4-a29272ff0d1f" (UID: "062b8e49-6137-4b89-8ae4-a29272ff0d1f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.619343 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/062b8e49-6137-4b89-8ae4-a29272ff0d1f-kube-api-access-z2szg" (OuterVolumeSpecName: "kube-api-access-z2szg") pod "062b8e49-6137-4b89-8ae4-a29272ff0d1f" (UID: "062b8e49-6137-4b89-8ae4-a29272ff0d1f"). InnerVolumeSpecName "kube-api-access-z2szg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.642538 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "062b8e49-6137-4b89-8ae4-a29272ff0d1f" (UID: "062b8e49-6137-4b89-8ae4-a29272ff0d1f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.644332 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-inventory" (OuterVolumeSpecName: "inventory") pod "062b8e49-6137-4b89-8ae4-a29272ff0d1f" (UID: "062b8e49-6137-4b89-8ae4-a29272ff0d1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.716388 4696 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.716432 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.716443 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2szg\" (UniqueName: \"kubernetes.io/projected/062b8e49-6137-4b89-8ae4-a29272ff0d1f-kube-api-access-z2szg\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:06 crc kubenswrapper[4696]: I1125 11:05:06.716453 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/062b8e49-6137-4b89-8ae4-a29272ff0d1f-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.133866 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" event={"ID":"062b8e49-6137-4b89-8ae4-a29272ff0d1f","Type":"ContainerDied","Data":"a06bd65c33cabce312aad4230c6e2de2e4ae50f4c7c3d9b38ff7ffbfdb38a49a"} Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.133910 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a06bd65c33cabce312aad4230c6e2de2e4ae50f4c7c3d9b38ff7ffbfdb38a49a" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.134248 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-65952" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.228605 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5"] Nov 25 11:05:07 crc kubenswrapper[4696]: E1125 11:05:07.229098 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a3da9a0-6262-4ff8-b510-e476db9d9985" containerName="keystone-cron" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.229125 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a3da9a0-6262-4ff8-b510-e476db9d9985" containerName="keystone-cron" Nov 25 11:05:07 crc kubenswrapper[4696]: E1125 11:05:07.229146 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062b8e49-6137-4b89-8ae4-a29272ff0d1f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.229156 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="062b8e49-6137-4b89-8ae4-a29272ff0d1f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.229404 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a3da9a0-6262-4ff8-b510-e476db9d9985" containerName="keystone-cron" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.229430 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="062b8e49-6137-4b89-8ae4-a29272ff0d1f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.230159 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.234896 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.234912 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.235028 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.235461 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.250612 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5"] Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.330749 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.330979 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjwwr\" (UniqueName: \"kubernetes.io/projected/38409cac-958b-497e-9431-4d0cafe2fd8d-kube-api-access-qjwwr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.331053 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.433084 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.433202 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjwwr\" (UniqueName: \"kubernetes.io/projected/38409cac-958b-497e-9431-4d0cafe2fd8d-kube-api-access-qjwwr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.433247 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.437483 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.437625 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.451898 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjwwr\" (UniqueName: \"kubernetes.io/projected/38409cac-958b-497e-9431-4d0cafe2fd8d-kube-api-access-qjwwr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-j29t5\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:07 crc kubenswrapper[4696]: I1125 11:05:07.552880 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:05:08 crc kubenswrapper[4696]: I1125 11:05:08.113512 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5"] Nov 25 11:05:08 crc kubenswrapper[4696]: I1125 11:05:08.121543 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:05:08 crc kubenswrapper[4696]: I1125 11:05:08.169018 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" event={"ID":"38409cac-958b-497e-9431-4d0cafe2fd8d","Type":"ContainerStarted","Data":"b2b791044de689174817db2e2004f3026821192108786f225db93d1c19406585"} Nov 25 11:05:09 crc kubenswrapper[4696]: I1125 11:05:09.178901 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" event={"ID":"38409cac-958b-497e-9431-4d0cafe2fd8d","Type":"ContainerStarted","Data":"7acf5b4c66db90c43453fa2139c88a384f9c71449e8d2c7a5ffac12b80848528"} Nov 25 11:05:09 crc kubenswrapper[4696]: I1125 11:05:09.204023 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" podStartSLOduration=1.8160690339999999 podStartE2EDuration="2.203998267s" podCreationTimestamp="2025-11-25 11:05:07 +0000 UTC" firstStartedPulling="2025-11-25 11:05:08.121308179 +0000 UTC m=+1704.934924766" lastFinishedPulling="2025-11-25 11:05:08.509237392 +0000 UTC m=+1705.322853999" observedRunningTime="2025-11-25 11:05:09.196501562 +0000 UTC m=+1706.010118149" watchObservedRunningTime="2025-11-25 11:05:09.203998267 +0000 UTC m=+1706.017614874" Nov 25 11:05:17 crc kubenswrapper[4696]: I1125 11:05:17.042570 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:05:17 crc kubenswrapper[4696]: E1125 11:05:17.043341 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:05:28 crc kubenswrapper[4696]: I1125 11:05:28.043436 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:05:28 crc kubenswrapper[4696]: E1125 11:05:28.044221 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:05:35 crc kubenswrapper[4696]: I1125 11:05:35.160200 4696 scope.go:117] "RemoveContainer" containerID="62be1a71b1f817ac50a79718ab71a3fa8ac78f3bd0933da18172e94b2d922df4" Nov 25 11:05:39 crc kubenswrapper[4696]: I1125 11:05:39.047481 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dq4dk"] Nov 25 11:05:39 crc kubenswrapper[4696]: I1125 11:05:39.055942 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-c2g2r"] Nov 25 11:05:39 crc kubenswrapper[4696]: I1125 11:05:39.065050 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dq4dk"] Nov 25 11:05:39 crc kubenswrapper[4696]: I1125 11:05:39.074766 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-c2g2r"] Nov 25 11:05:40 crc kubenswrapper[4696]: I1125 11:05:40.054057 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1bfd7e7-4caa-4b79-be66-004376b3f7f4" path="/var/lib/kubelet/pods/c1bfd7e7-4caa-4b79-be66-004376b3f7f4/volumes" Nov 25 11:05:40 crc kubenswrapper[4696]: I1125 11:05:40.071379 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6" path="/var/lib/kubelet/pods/e5c18c5a-6cef-4afe-a3a1-f0ad04cec1a6/volumes" Nov 25 11:05:43 crc kubenswrapper[4696]: I1125 11:05:43.042777 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:05:43 crc kubenswrapper[4696]: E1125 11:05:43.043462 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:05:51 crc kubenswrapper[4696]: I1125 11:05:51.057241 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nqj4p"] Nov 25 11:05:51 crc kubenswrapper[4696]: I1125 11:05:51.070495 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5vqmj"] Nov 25 11:05:51 crc kubenswrapper[4696]: I1125 11:05:51.087588 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nqj4p"] Nov 25 11:05:51 crc kubenswrapper[4696]: I1125 11:05:51.095822 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5vqmj"] Nov 25 11:05:52 crc kubenswrapper[4696]: I1125 11:05:52.073952 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a052a79-e780-430c-a049-cc01418ba194" path="/var/lib/kubelet/pods/3a052a79-e780-430c-a049-cc01418ba194/volumes" Nov 25 11:05:52 crc kubenswrapper[4696]: I1125 11:05:52.077279 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cacd4fa8-b70a-451d-92a9-40e53ea0a406" path="/var/lib/kubelet/pods/cacd4fa8-b70a-451d-92a9-40e53ea0a406/volumes" Nov 25 11:05:55 crc kubenswrapper[4696]: I1125 11:05:55.042310 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:05:55 crc kubenswrapper[4696]: E1125 11:05:55.042880 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:05:59 crc kubenswrapper[4696]: I1125 11:05:59.031089 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7qtsq"] Nov 25 11:05:59 crc kubenswrapper[4696]: I1125 11:05:59.038811 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7qtsq"] Nov 25 11:06:00 crc kubenswrapper[4696]: I1125 11:06:00.058596 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9353d93a-d886-43d0-a4d4-6ad17a28f52a" path="/var/lib/kubelet/pods/9353d93a-d886-43d0-a4d4-6ad17a28f52a/volumes" Nov 25 11:06:10 crc kubenswrapper[4696]: I1125 11:06:10.042473 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:06:10 crc kubenswrapper[4696]: E1125 11:06:10.043249 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:06:24 crc kubenswrapper[4696]: I1125 11:06:24.048237 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:06:24 crc kubenswrapper[4696]: E1125 11:06:24.049065 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:06:35 crc kubenswrapper[4696]: I1125 11:06:35.242253 4696 scope.go:117] "RemoveContainer" containerID="0e534c69c8ef2d7aa45777809c98dc906cf2231b90e8f435f7592bb0301cef72" Nov 25 11:06:35 crc kubenswrapper[4696]: I1125 11:06:35.338827 4696 scope.go:117] "RemoveContainer" containerID="aea37d144e9450979ed996b37ced91a9f9660ed6c748e718a17f0296d80bb5ad" Nov 25 11:06:35 crc kubenswrapper[4696]: I1125 11:06:35.368474 4696 scope.go:117] "RemoveContainer" containerID="5c97bb9514113109c93a46209e8213df0786b15a142aff872fbb81cc5256e3a0" Nov 25 11:06:35 crc kubenswrapper[4696]: I1125 11:06:35.444394 4696 scope.go:117] "RemoveContainer" containerID="00f8873d95d129a119835fee128c27e00b740b6dcaadce41e01b6bdcd3f20925" Nov 25 11:06:35 crc kubenswrapper[4696]: I1125 11:06:35.472154 4696 scope.go:117] "RemoveContainer" containerID="efac50fa2c171a3ffac9514f2b79a3f7d9e722213d7a45a4717d5568fbb2cabf" Nov 25 11:06:37 crc kubenswrapper[4696]: I1125 11:06:37.043627 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:06:37 crc kubenswrapper[4696]: E1125 11:06:37.044488 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.068033 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-162c-account-create-sn7k7"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.090925 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-x9l8x"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.112738 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-2731-account-create-tprt5"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.123656 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-162c-account-create-sn7k7"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.134972 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-x9l8x"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.150760 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ac60-account-create-grcfr"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.166785 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-2731-account-create-tprt5"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.176026 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hdlx8"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.184755 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-plnjv"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.192883 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ac60-account-create-grcfr"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.200383 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-plnjv"] Nov 25 11:06:44 crc kubenswrapper[4696]: I1125 11:06:44.209149 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hdlx8"] Nov 25 11:06:46 crc kubenswrapper[4696]: I1125 11:06:46.053980 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ba9c13d-1506-46fd-822e-4d816ecff490" path="/var/lib/kubelet/pods/2ba9c13d-1506-46fd-822e-4d816ecff490/volumes" Nov 25 11:06:46 crc kubenswrapper[4696]: I1125 11:06:46.055278 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636b96ee-a180-4234-9a9b-f1ebfd9a2f25" path="/var/lib/kubelet/pods/636b96ee-a180-4234-9a9b-f1ebfd9a2f25/volumes" Nov 25 11:06:46 crc kubenswrapper[4696]: I1125 11:06:46.056135 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95c4ea9f-5c8c-4475-9c5c-3c00ef595f19" path="/var/lib/kubelet/pods/95c4ea9f-5c8c-4475-9c5c-3c00ef595f19/volumes" Nov 25 11:06:46 crc kubenswrapper[4696]: I1125 11:06:46.057053 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b183e265-9749-45f5-9b87-de239a7daca0" path="/var/lib/kubelet/pods/b183e265-9749-45f5-9b87-de239a7daca0/volumes" Nov 25 11:06:46 crc kubenswrapper[4696]: I1125 11:06:46.058669 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b276d474-ccfa-4b14-95e6-a92d68897387" path="/var/lib/kubelet/pods/b276d474-ccfa-4b14-95e6-a92d68897387/volumes" Nov 25 11:06:46 crc kubenswrapper[4696]: I1125 11:06:46.059647 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0" path="/var/lib/kubelet/pods/bf4e7bd1-4e75-4ed5-a169-ae36918ca2a0/volumes" Nov 25 11:06:49 crc kubenswrapper[4696]: I1125 11:06:49.042949 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:06:49 crc kubenswrapper[4696]: E1125 11:06:49.044172 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:07:00 crc kubenswrapper[4696]: I1125 11:07:00.042162 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:07:00 crc kubenswrapper[4696]: E1125 11:07:00.042835 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:07:13 crc kubenswrapper[4696]: I1125 11:07:13.043971 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:07:13 crc kubenswrapper[4696]: E1125 11:07:13.046430 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:07:20 crc kubenswrapper[4696]: I1125 11:07:20.061057 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7w8x2"] Nov 25 11:07:20 crc kubenswrapper[4696]: I1125 11:07:20.073808 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7w8x2"] Nov 25 11:07:22 crc kubenswrapper[4696]: I1125 11:07:22.054098 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ce5f08-60ba-44bb-865c-b01c57ae49ac" path="/var/lib/kubelet/pods/35ce5f08-60ba-44bb-865c-b01c57ae49ac/volumes" Nov 25 11:07:25 crc kubenswrapper[4696]: I1125 11:07:25.043699 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:07:25 crc kubenswrapper[4696]: E1125 11:07:25.044533 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.755760 4696 scope.go:117] "RemoveContainer" containerID="541f33e91f124886097bff5992fb8bbec7619b894e803294c7fd06e12c9c5e6f" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.786275 4696 scope.go:117] "RemoveContainer" containerID="3c7adebf28f9d5cf798a7e088bc7cc4ca91ecfcdc49dbb27cb1727dcee908ef5" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.828835 4696 scope.go:117] "RemoveContainer" containerID="dcb0238dd2109f37a083c8a5e09012b1ee582d563e8545ca7fc0a57b00729346" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.863264 4696 scope.go:117] "RemoveContainer" containerID="b067cca4b293436a9051abdb47d1fa1c8ce3689f242443c36f855fb3b5b27d3d" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.908578 4696 scope.go:117] "RemoveContainer" containerID="704497debdd2b3d4266c02940295a55717e3f4ded974697b047eccb74c8eeb1e" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.954867 4696 scope.go:117] "RemoveContainer" containerID="1c344fc7b37117bd87c2080daf5416fc850b91049a7cd609bac58ee740817368" Nov 25 11:07:35 crc kubenswrapper[4696]: I1125 11:07:35.997124 4696 scope.go:117] "RemoveContainer" containerID="49b5f5810bc0b4b182f8276f34fca35c5498ba49f506a497888c582525a02b21" Nov 25 11:07:39 crc kubenswrapper[4696]: I1125 11:07:39.036104 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfvg7"] Nov 25 11:07:39 crc kubenswrapper[4696]: I1125 11:07:39.044001 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lfvg7"] Nov 25 11:07:40 crc kubenswrapper[4696]: I1125 11:07:40.043367 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:07:40 crc kubenswrapper[4696]: E1125 11:07:40.043624 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:07:40 crc kubenswrapper[4696]: I1125 11:07:40.054917 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef638cd7-5ef4-4c2e-96e9-d7655bc6f527" path="/var/lib/kubelet/pods/ef638cd7-5ef4-4c2e-96e9-d7655bc6f527/volumes" Nov 25 11:07:44 crc kubenswrapper[4696]: I1125 11:07:44.062546 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfdng"] Nov 25 11:07:44 crc kubenswrapper[4696]: I1125 11:07:44.090062 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xfdng"] Nov 25 11:07:46 crc kubenswrapper[4696]: I1125 11:07:46.053135 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b10ca5a-c03c-407e-9d19-c649df1f9711" path="/var/lib/kubelet/pods/4b10ca5a-c03c-407e-9d19-c649df1f9711/volumes" Nov 25 11:07:52 crc kubenswrapper[4696]: I1125 11:07:52.612390 4696 generic.go:334] "Generic (PLEG): container finished" podID="38409cac-958b-497e-9431-4d0cafe2fd8d" containerID="7acf5b4c66db90c43453fa2139c88a384f9c71449e8d2c7a5ffac12b80848528" exitCode=0 Nov 25 11:07:52 crc kubenswrapper[4696]: I1125 11:07:52.612471 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" event={"ID":"38409cac-958b-497e-9431-4d0cafe2fd8d","Type":"ContainerDied","Data":"7acf5b4c66db90c43453fa2139c88a384f9c71449e8d2c7a5ffac12b80848528"} Nov 25 11:07:53 crc kubenswrapper[4696]: I1125 11:07:53.043023 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:07:53 crc kubenswrapper[4696]: E1125 11:07:53.043585 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.093832 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.225553 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjwwr\" (UniqueName: \"kubernetes.io/projected/38409cac-958b-497e-9431-4d0cafe2fd8d-kube-api-access-qjwwr\") pod \"38409cac-958b-497e-9431-4d0cafe2fd8d\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.225693 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-inventory\") pod \"38409cac-958b-497e-9431-4d0cafe2fd8d\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.225737 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-ssh-key\") pod \"38409cac-958b-497e-9431-4d0cafe2fd8d\" (UID: \"38409cac-958b-497e-9431-4d0cafe2fd8d\") " Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.234699 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38409cac-958b-497e-9431-4d0cafe2fd8d-kube-api-access-qjwwr" (OuterVolumeSpecName: "kube-api-access-qjwwr") pod "38409cac-958b-497e-9431-4d0cafe2fd8d" (UID: "38409cac-958b-497e-9431-4d0cafe2fd8d"). InnerVolumeSpecName "kube-api-access-qjwwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.257418 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "38409cac-958b-497e-9431-4d0cafe2fd8d" (UID: "38409cac-958b-497e-9431-4d0cafe2fd8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.257449 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-inventory" (OuterVolumeSpecName: "inventory") pod "38409cac-958b-497e-9431-4d0cafe2fd8d" (UID: "38409cac-958b-497e-9431-4d0cafe2fd8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.328772 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjwwr\" (UniqueName: \"kubernetes.io/projected/38409cac-958b-497e-9431-4d0cafe2fd8d-kube-api-access-qjwwr\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.328820 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.328831 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38409cac-958b-497e-9431-4d0cafe2fd8d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.631920 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" event={"ID":"38409cac-958b-497e-9431-4d0cafe2fd8d","Type":"ContainerDied","Data":"b2b791044de689174817db2e2004f3026821192108786f225db93d1c19406585"} Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.631961 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2b791044de689174817db2e2004f3026821192108786f225db93d1c19406585" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.631980 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-j29t5" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.738160 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp"] Nov 25 11:07:54 crc kubenswrapper[4696]: E1125 11:07:54.738594 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38409cac-958b-497e-9431-4d0cafe2fd8d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.738607 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38409cac-958b-497e-9431-4d0cafe2fd8d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.738813 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="38409cac-958b-497e-9431-4d0cafe2fd8d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.739442 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.745886 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.746212 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.746394 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.746537 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.754032 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp"] Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.937746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbg5k\" (UniqueName: \"kubernetes.io/projected/574f645b-d7d9-4756-82ef-27bfa7bf982e-kube-api-access-dbg5k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.937867 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:54 crc kubenswrapper[4696]: I1125 11:07:54.938007 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.039421 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.039471 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.039592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbg5k\" (UniqueName: \"kubernetes.io/projected/574f645b-d7d9-4756-82ef-27bfa7bf982e-kube-api-access-dbg5k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.045016 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.048701 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.064473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbg5k\" (UniqueName: \"kubernetes.io/projected/574f645b-d7d9-4756-82ef-27bfa7bf982e-kube-api-access-dbg5k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.066609 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:07:55 crc kubenswrapper[4696]: I1125 11:07:55.750880 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp"] Nov 25 11:07:56 crc kubenswrapper[4696]: I1125 11:07:56.648454 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" event={"ID":"574f645b-d7d9-4756-82ef-27bfa7bf982e","Type":"ContainerStarted","Data":"6d0d3b0a774bb44a0420b4bf9b08418c13977e850ba4318ab2a6ba48e1d42653"} Nov 25 11:07:56 crc kubenswrapper[4696]: I1125 11:07:56.648835 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" event={"ID":"574f645b-d7d9-4756-82ef-27bfa7bf982e","Type":"ContainerStarted","Data":"fb36826294a82f2ea3bdf26932de1f7ffc41860336a3caf79915e1db388c8a87"} Nov 25 11:07:56 crc kubenswrapper[4696]: I1125 11:07:56.665443 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" podStartSLOduration=2.1829116490000002 podStartE2EDuration="2.665419192s" podCreationTimestamp="2025-11-25 11:07:54 +0000 UTC" firstStartedPulling="2025-11-25 11:07:55.756554484 +0000 UTC m=+1872.570171071" lastFinishedPulling="2025-11-25 11:07:56.239062027 +0000 UTC m=+1873.052678614" observedRunningTime="2025-11-25 11:07:56.662718867 +0000 UTC m=+1873.476335464" watchObservedRunningTime="2025-11-25 11:07:56.665419192 +0000 UTC m=+1873.479035779" Nov 25 11:08:06 crc kubenswrapper[4696]: I1125 11:08:06.042259 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:08:06 crc kubenswrapper[4696]: E1125 11:08:06.043020 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:08:17 crc kubenswrapper[4696]: I1125 11:08:17.044075 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:08:17 crc kubenswrapper[4696]: E1125 11:08:17.045692 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:08:23 crc kubenswrapper[4696]: I1125 11:08:23.039846 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2s8v9"] Nov 25 11:08:23 crc kubenswrapper[4696]: I1125 11:08:23.051314 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2s8v9"] Nov 25 11:08:24 crc kubenswrapper[4696]: I1125 11:08:24.052582 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06860896-de50-41b9-9770-156967ac658f" path="/var/lib/kubelet/pods/06860896-de50-41b9-9770-156967ac658f/volumes" Nov 25 11:08:29 crc kubenswrapper[4696]: I1125 11:08:29.042322 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:08:29 crc kubenswrapper[4696]: E1125 11:08:29.043139 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:08:36 crc kubenswrapper[4696]: I1125 11:08:36.139266 4696 scope.go:117] "RemoveContainer" containerID="96e7cd861737e0e9581db9df692210348ae8c3c1c31ea2adea0c671058b1f2c5" Nov 25 11:08:36 crc kubenswrapper[4696]: I1125 11:08:36.176752 4696 scope.go:117] "RemoveContainer" containerID="4d316d2a2a1f34711da8f0a58a99152d05248979396e27f75374e53e5f17b288" Nov 25 11:08:36 crc kubenswrapper[4696]: I1125 11:08:36.214141 4696 scope.go:117] "RemoveContainer" containerID="763a7b7a7e164083e8a32a3a178cea38e62790734ded56d23a26b1b113345c41" Nov 25 11:08:40 crc kubenswrapper[4696]: I1125 11:08:40.042192 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:08:40 crc kubenswrapper[4696]: E1125 11:08:40.043148 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:08:53 crc kubenswrapper[4696]: I1125 11:08:53.042343 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:08:53 crc kubenswrapper[4696]: E1125 11:08:53.044179 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:09:08 crc kubenswrapper[4696]: I1125 11:09:08.042566 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:09:08 crc kubenswrapper[4696]: E1125 11:09:08.044261 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:09:18 crc kubenswrapper[4696]: I1125 11:09:18.305811 4696 generic.go:334] "Generic (PLEG): container finished" podID="574f645b-d7d9-4756-82ef-27bfa7bf982e" containerID="6d0d3b0a774bb44a0420b4bf9b08418c13977e850ba4318ab2a6ba48e1d42653" exitCode=0 Nov 25 11:09:18 crc kubenswrapper[4696]: I1125 11:09:18.305881 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" event={"ID":"574f645b-d7d9-4756-82ef-27bfa7bf982e","Type":"ContainerDied","Data":"6d0d3b0a774bb44a0420b4bf9b08418c13977e850ba4318ab2a6ba48e1d42653"} Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.043065 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:09:19 crc kubenswrapper[4696]: E1125 11:09:19.043609 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.716370 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.873955 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-inventory\") pod \"574f645b-d7d9-4756-82ef-27bfa7bf982e\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.874046 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbg5k\" (UniqueName: \"kubernetes.io/projected/574f645b-d7d9-4756-82ef-27bfa7bf982e-kube-api-access-dbg5k\") pod \"574f645b-d7d9-4756-82ef-27bfa7bf982e\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.874104 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-ssh-key\") pod \"574f645b-d7d9-4756-82ef-27bfa7bf982e\" (UID: \"574f645b-d7d9-4756-82ef-27bfa7bf982e\") " Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.879912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574f645b-d7d9-4756-82ef-27bfa7bf982e-kube-api-access-dbg5k" (OuterVolumeSpecName: "kube-api-access-dbg5k") pod "574f645b-d7d9-4756-82ef-27bfa7bf982e" (UID: "574f645b-d7d9-4756-82ef-27bfa7bf982e"). InnerVolumeSpecName "kube-api-access-dbg5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.907804 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-inventory" (OuterVolumeSpecName: "inventory") pod "574f645b-d7d9-4756-82ef-27bfa7bf982e" (UID: "574f645b-d7d9-4756-82ef-27bfa7bf982e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.910742 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "574f645b-d7d9-4756-82ef-27bfa7bf982e" (UID: "574f645b-d7d9-4756-82ef-27bfa7bf982e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.976111 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.976176 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/574f645b-d7d9-4756-82ef-27bfa7bf982e-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:19 crc kubenswrapper[4696]: I1125 11:09:19.976188 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbg5k\" (UniqueName: \"kubernetes.io/projected/574f645b-d7d9-4756-82ef-27bfa7bf982e-kube-api-access-dbg5k\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.324967 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" event={"ID":"574f645b-d7d9-4756-82ef-27bfa7bf982e","Type":"ContainerDied","Data":"fb36826294a82f2ea3bdf26932de1f7ffc41860336a3caf79915e1db388c8a87"} Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.325011 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb36826294a82f2ea3bdf26932de1f7ffc41860336a3caf79915e1db388c8a87" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.325031 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.469490 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s"] Nov 25 11:09:20 crc kubenswrapper[4696]: E1125 11:09:20.469900 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574f645b-d7d9-4756-82ef-27bfa7bf982e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.469917 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="574f645b-d7d9-4756-82ef-27bfa7bf982e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.470107 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="574f645b-d7d9-4756-82ef-27bfa7bf982e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.470758 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.473349 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.474103 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.475323 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.479834 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s"] Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.479872 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.585688 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.585771 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhflm\" (UniqueName: \"kubernetes.io/projected/b756844d-cdfd-4345-8a32-035b71161bfc-kube-api-access-rhflm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.585866 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.687806 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.687869 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhflm\" (UniqueName: \"kubernetes.io/projected/b756844d-cdfd-4345-8a32-035b71161bfc-kube-api-access-rhflm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.687932 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.693176 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.695949 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.713069 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhflm\" (UniqueName: \"kubernetes.io/projected/b756844d-cdfd-4345-8a32-035b71161bfc-kube-api-access-rhflm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:20 crc kubenswrapper[4696]: I1125 11:09:20.786006 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:21 crc kubenswrapper[4696]: I1125 11:09:21.296611 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s"] Nov 25 11:09:21 crc kubenswrapper[4696]: I1125 11:09:21.338688 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" event={"ID":"b756844d-cdfd-4345-8a32-035b71161bfc","Type":"ContainerStarted","Data":"3e0df0a245729be681d183f40c8d2364719e5ab704237d5244a0fc50af34c64d"} Nov 25 11:09:22 crc kubenswrapper[4696]: I1125 11:09:22.350434 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" event={"ID":"b756844d-cdfd-4345-8a32-035b71161bfc","Type":"ContainerStarted","Data":"dc66f138de6dc5740a93073ba25b863398bb681784e8f066b75e03506f52039b"} Nov 25 11:09:22 crc kubenswrapper[4696]: I1125 11:09:22.369676 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" podStartSLOduration=1.768910162 podStartE2EDuration="2.369636013s" podCreationTimestamp="2025-11-25 11:09:20 +0000 UTC" firstStartedPulling="2025-11-25 11:09:21.300721755 +0000 UTC m=+1958.114338342" lastFinishedPulling="2025-11-25 11:09:21.901447606 +0000 UTC m=+1958.715064193" observedRunningTime="2025-11-25 11:09:22.365723049 +0000 UTC m=+1959.179339656" watchObservedRunningTime="2025-11-25 11:09:22.369636013 +0000 UTC m=+1959.183252600" Nov 25 11:09:28 crc kubenswrapper[4696]: I1125 11:09:28.407525 4696 generic.go:334] "Generic (PLEG): container finished" podID="b756844d-cdfd-4345-8a32-035b71161bfc" containerID="dc66f138de6dc5740a93073ba25b863398bb681784e8f066b75e03506f52039b" exitCode=0 Nov 25 11:09:28 crc kubenswrapper[4696]: I1125 11:09:28.407594 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" event={"ID":"b756844d-cdfd-4345-8a32-035b71161bfc","Type":"ContainerDied","Data":"dc66f138de6dc5740a93073ba25b863398bb681784e8f066b75e03506f52039b"} Nov 25 11:09:29 crc kubenswrapper[4696]: I1125 11:09:29.839205 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:29 crc kubenswrapper[4696]: I1125 11:09:29.967026 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhflm\" (UniqueName: \"kubernetes.io/projected/b756844d-cdfd-4345-8a32-035b71161bfc-kube-api-access-rhflm\") pod \"b756844d-cdfd-4345-8a32-035b71161bfc\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " Nov 25 11:09:29 crc kubenswrapper[4696]: I1125 11:09:29.967443 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-ssh-key\") pod \"b756844d-cdfd-4345-8a32-035b71161bfc\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " Nov 25 11:09:29 crc kubenswrapper[4696]: I1125 11:09:29.967533 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-inventory\") pod \"b756844d-cdfd-4345-8a32-035b71161bfc\" (UID: \"b756844d-cdfd-4345-8a32-035b71161bfc\") " Nov 25 11:09:29 crc kubenswrapper[4696]: I1125 11:09:29.972214 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b756844d-cdfd-4345-8a32-035b71161bfc-kube-api-access-rhflm" (OuterVolumeSpecName: "kube-api-access-rhflm") pod "b756844d-cdfd-4345-8a32-035b71161bfc" (UID: "b756844d-cdfd-4345-8a32-035b71161bfc"). InnerVolumeSpecName "kube-api-access-rhflm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:09:29 crc kubenswrapper[4696]: I1125 11:09:29.993655 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b756844d-cdfd-4345-8a32-035b71161bfc" (UID: "b756844d-cdfd-4345-8a32-035b71161bfc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.002517 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-inventory" (OuterVolumeSpecName: "inventory") pod "b756844d-cdfd-4345-8a32-035b71161bfc" (UID: "b756844d-cdfd-4345-8a32-035b71161bfc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.069366 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.069402 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b756844d-cdfd-4345-8a32-035b71161bfc-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.069414 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhflm\" (UniqueName: \"kubernetes.io/projected/b756844d-cdfd-4345-8a32-035b71161bfc-kube-api-access-rhflm\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:30 crc kubenswrapper[4696]: E1125 11:09:30.226226 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb756844d_cdfd_4345_8a32_035b71161bfc.slice\": RecentStats: unable to find data in memory cache]" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.423775 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" event={"ID":"b756844d-cdfd-4345-8a32-035b71161bfc","Type":"ContainerDied","Data":"3e0df0a245729be681d183f40c8d2364719e5ab704237d5244a0fc50af34c64d"} Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.424261 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e0df0a245729be681d183f40c8d2364719e5ab704237d5244a0fc50af34c64d" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.423842 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.496648 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5"] Nov 25 11:09:30 crc kubenswrapper[4696]: E1125 11:09:30.497105 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b756844d-cdfd-4345-8a32-035b71161bfc" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.497129 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b756844d-cdfd-4345-8a32-035b71161bfc" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.497360 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b756844d-cdfd-4345-8a32-035b71161bfc" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.498341 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.500779 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.500837 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.501351 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.501379 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.507216 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5"] Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.582845 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7g92\" (UniqueName: \"kubernetes.io/projected/d69df5c8-e28a-4b73-9d2b-97087aaed351-kube-api-access-r7g92\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.583048 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.583351 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.685124 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.685206 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.685266 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7g92\" (UniqueName: \"kubernetes.io/projected/d69df5c8-e28a-4b73-9d2b-97087aaed351-kube-api-access-r7g92\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.688994 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.689555 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.705029 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7g92\" (UniqueName: \"kubernetes.io/projected/d69df5c8-e28a-4b73-9d2b-97087aaed351-kube-api-access-r7g92\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kg6v5\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:30 crc kubenswrapper[4696]: I1125 11:09:30.832894 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:09:31 crc kubenswrapper[4696]: I1125 11:09:31.042933 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:09:31 crc kubenswrapper[4696]: E1125 11:09:31.043735 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:09:31 crc kubenswrapper[4696]: I1125 11:09:31.303653 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5"] Nov 25 11:09:31 crc kubenswrapper[4696]: I1125 11:09:31.436065 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" event={"ID":"d69df5c8-e28a-4b73-9d2b-97087aaed351","Type":"ContainerStarted","Data":"78927f7da77d3598b3d27500f265bcc6e645553354d1efc15e0b97a5941ea875"} Nov 25 11:09:32 crc kubenswrapper[4696]: I1125 11:09:32.446493 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" event={"ID":"d69df5c8-e28a-4b73-9d2b-97087aaed351","Type":"ContainerStarted","Data":"a9f90fda84ace6e3c775b1c7dbb71a1ce436f10bcc2a3202ee342403f514fa36"} Nov 25 11:09:32 crc kubenswrapper[4696]: I1125 11:09:32.470042 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" podStartSLOduration=2.075170771 podStartE2EDuration="2.470018327s" podCreationTimestamp="2025-11-25 11:09:30 +0000 UTC" firstStartedPulling="2025-11-25 11:09:31.304741739 +0000 UTC m=+1968.118358326" lastFinishedPulling="2025-11-25 11:09:31.699589295 +0000 UTC m=+1968.513205882" observedRunningTime="2025-11-25 11:09:32.466738563 +0000 UTC m=+1969.280355170" watchObservedRunningTime="2025-11-25 11:09:32.470018327 +0000 UTC m=+1969.283634914" Nov 25 11:09:42 crc kubenswrapper[4696]: I1125 11:09:42.043089 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:09:42 crc kubenswrapper[4696]: E1125 11:09:42.044108 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.737292 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6jz7q"] Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.740122 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.761641 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jz7q"] Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.884865 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-catalog-content\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.885120 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-utilities\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.885254 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-966fp\" (UniqueName: \"kubernetes.io/projected/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-kube-api-access-966fp\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.986600 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-catalog-content\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.986689 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-utilities\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.986732 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-966fp\" (UniqueName: \"kubernetes.io/projected/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-kube-api-access-966fp\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.987163 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-catalog-content\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:51 crc kubenswrapper[4696]: I1125 11:09:51.987209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-utilities\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:52 crc kubenswrapper[4696]: I1125 11:09:52.020506 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-966fp\" (UniqueName: \"kubernetes.io/projected/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-kube-api-access-966fp\") pod \"redhat-marketplace-6jz7q\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:52 crc kubenswrapper[4696]: I1125 11:09:52.062247 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:09:52 crc kubenswrapper[4696]: I1125 11:09:52.541653 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jz7q"] Nov 25 11:09:52 crc kubenswrapper[4696]: W1125 11:09:52.548818 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb199b23f_df6a_409f_a6fc_fe60ecb4bb2f.slice/crio-1790bcf70c5897201d16b618de5997ddd6e1723307734fd8850932580592cdc1 WatchSource:0}: Error finding container 1790bcf70c5897201d16b618de5997ddd6e1723307734fd8850932580592cdc1: Status 404 returned error can't find the container with id 1790bcf70c5897201d16b618de5997ddd6e1723307734fd8850932580592cdc1 Nov 25 11:09:52 crc kubenswrapper[4696]: I1125 11:09:52.605122 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jz7q" event={"ID":"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f","Type":"ContainerStarted","Data":"1790bcf70c5897201d16b618de5997ddd6e1723307734fd8850932580592cdc1"} Nov 25 11:09:53 crc kubenswrapper[4696]: I1125 11:09:53.612703 4696 generic.go:334] "Generic (PLEG): container finished" podID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerID="be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c" exitCode=0 Nov 25 11:09:53 crc kubenswrapper[4696]: I1125 11:09:53.612782 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jz7q" event={"ID":"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f","Type":"ContainerDied","Data":"be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c"} Nov 25 11:09:55 crc kubenswrapper[4696]: I1125 11:09:55.041744 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:09:55 crc kubenswrapper[4696]: E1125 11:09:55.042274 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.131160 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pt2hf"] Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.134788 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.167740 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pt2hf"] Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.279722 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-utilities\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.280205 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhcm6\" (UniqueName: \"kubernetes.io/projected/bba18af9-27fa-4993-99a7-d397563e0aa5-kube-api-access-fhcm6\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.280338 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-catalog-content\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.382367 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-catalog-content\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.382541 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-utilities\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.382625 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhcm6\" (UniqueName: \"kubernetes.io/projected/bba18af9-27fa-4993-99a7-d397563e0aa5-kube-api-access-fhcm6\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.383062 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-catalog-content\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.383313 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-utilities\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.408553 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhcm6\" (UniqueName: \"kubernetes.io/projected/bba18af9-27fa-4993-99a7-d397563e0aa5-kube-api-access-fhcm6\") pod \"community-operators-pt2hf\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.463729 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.649453 4696 generic.go:334] "Generic (PLEG): container finished" podID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerID="58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52" exitCode=0 Nov 25 11:09:56 crc kubenswrapper[4696]: I1125 11:09:56.650520 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jz7q" event={"ID":"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f","Type":"ContainerDied","Data":"58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52"} Nov 25 11:09:57 crc kubenswrapper[4696]: I1125 11:09:57.044403 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pt2hf"] Nov 25 11:09:57 crc kubenswrapper[4696]: W1125 11:09:57.047462 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbba18af9_27fa_4993_99a7_d397563e0aa5.slice/crio-6c5d6a6f265731775c4748a0f210ba595d9bb1e68e02eac9f7d11e9b2f64a01e WatchSource:0}: Error finding container 6c5d6a6f265731775c4748a0f210ba595d9bb1e68e02eac9f7d11e9b2f64a01e: Status 404 returned error can't find the container with id 6c5d6a6f265731775c4748a0f210ba595d9bb1e68e02eac9f7d11e9b2f64a01e Nov 25 11:09:57 crc kubenswrapper[4696]: I1125 11:09:57.661369 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jz7q" event={"ID":"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f","Type":"ContainerStarted","Data":"525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8"} Nov 25 11:09:57 crc kubenswrapper[4696]: I1125 11:09:57.663943 4696 generic.go:334] "Generic (PLEG): container finished" podID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerID="7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306" exitCode=0 Nov 25 11:09:57 crc kubenswrapper[4696]: I1125 11:09:57.663977 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerDied","Data":"7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306"} Nov 25 11:09:57 crc kubenswrapper[4696]: I1125 11:09:57.663993 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerStarted","Data":"6c5d6a6f265731775c4748a0f210ba595d9bb1e68e02eac9f7d11e9b2f64a01e"} Nov 25 11:09:57 crc kubenswrapper[4696]: I1125 11:09:57.686118 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6jz7q" podStartSLOduration=3.167453504 podStartE2EDuration="6.686095502s" podCreationTimestamp="2025-11-25 11:09:51 +0000 UTC" firstStartedPulling="2025-11-25 11:09:53.61435417 +0000 UTC m=+1990.427970757" lastFinishedPulling="2025-11-25 11:09:57.132996168 +0000 UTC m=+1993.946612755" observedRunningTime="2025-11-25 11:09:57.682794397 +0000 UTC m=+1994.496411014" watchObservedRunningTime="2025-11-25 11:09:57.686095502 +0000 UTC m=+1994.499712099" Nov 25 11:09:58 crc kubenswrapper[4696]: I1125 11:09:58.676437 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerStarted","Data":"4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495"} Nov 25 11:10:01 crc kubenswrapper[4696]: I1125 11:10:01.700648 4696 generic.go:334] "Generic (PLEG): container finished" podID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerID="4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495" exitCode=0 Nov 25 11:10:01 crc kubenswrapper[4696]: I1125 11:10:01.700876 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerDied","Data":"4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495"} Nov 25 11:10:02 crc kubenswrapper[4696]: I1125 11:10:02.063057 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:10:02 crc kubenswrapper[4696]: I1125 11:10:02.063094 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:10:02 crc kubenswrapper[4696]: I1125 11:10:02.116641 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:10:02 crc kubenswrapper[4696]: I1125 11:10:02.711857 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerStarted","Data":"2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590"} Nov 25 11:10:02 crc kubenswrapper[4696]: I1125 11:10:02.744086 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pt2hf" podStartSLOduration=2.262819663 podStartE2EDuration="6.744063255s" podCreationTimestamp="2025-11-25 11:09:56 +0000 UTC" firstStartedPulling="2025-11-25 11:09:57.665711176 +0000 UTC m=+1994.479327763" lastFinishedPulling="2025-11-25 11:10:02.146954768 +0000 UTC m=+1998.960571355" observedRunningTime="2025-11-25 11:10:02.734079769 +0000 UTC m=+1999.547696366" watchObservedRunningTime="2025-11-25 11:10:02.744063255 +0000 UTC m=+1999.557679842" Nov 25 11:10:02 crc kubenswrapper[4696]: I1125 11:10:02.768051 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:10:03 crc kubenswrapper[4696]: I1125 11:10:03.709554 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jz7q"] Nov 25 11:10:04 crc kubenswrapper[4696]: I1125 11:10:04.728343 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6jz7q" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="registry-server" containerID="cri-o://525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8" gracePeriod=2 Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.298817 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.479051 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-utilities\") pod \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.479149 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-966fp\" (UniqueName: \"kubernetes.io/projected/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-kube-api-access-966fp\") pod \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.479311 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-catalog-content\") pod \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\" (UID: \"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f\") " Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.479866 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-utilities" (OuterVolumeSpecName: "utilities") pod "b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" (UID: "b199b23f-df6a-409f-a6fc-fe60ecb4bb2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.486796 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-kube-api-access-966fp" (OuterVolumeSpecName: "kube-api-access-966fp") pod "b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" (UID: "b199b23f-df6a-409f-a6fc-fe60ecb4bb2f"). InnerVolumeSpecName "kube-api-access-966fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.494789 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" (UID: "b199b23f-df6a-409f-a6fc-fe60ecb4bb2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.581092 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-966fp\" (UniqueName: \"kubernetes.io/projected/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-kube-api-access-966fp\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.581128 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.581137 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.739856 4696 generic.go:334] "Generic (PLEG): container finished" podID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerID="525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8" exitCode=0 Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.739911 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jz7q" event={"ID":"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f","Type":"ContainerDied","Data":"525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8"} Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.739934 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jz7q" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.739957 4696 scope.go:117] "RemoveContainer" containerID="525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.739943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jz7q" event={"ID":"b199b23f-df6a-409f-a6fc-fe60ecb4bb2f","Type":"ContainerDied","Data":"1790bcf70c5897201d16b618de5997ddd6e1723307734fd8850932580592cdc1"} Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.760422 4696 scope.go:117] "RemoveContainer" containerID="58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.779289 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jz7q"] Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.787407 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jz7q"] Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.802521 4696 scope.go:117] "RemoveContainer" containerID="be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.833071 4696 scope.go:117] "RemoveContainer" containerID="525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8" Nov 25 11:10:05 crc kubenswrapper[4696]: E1125 11:10:05.833555 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8\": container with ID starting with 525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8 not found: ID does not exist" containerID="525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.833591 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8"} err="failed to get container status \"525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8\": rpc error: code = NotFound desc = could not find container \"525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8\": container with ID starting with 525c886c5782eae964cb15c3437e822d0943e4fd008dad0c95e2b4a9a629fcc8 not found: ID does not exist" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.833612 4696 scope.go:117] "RemoveContainer" containerID="58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52" Nov 25 11:10:05 crc kubenswrapper[4696]: E1125 11:10:05.833931 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52\": container with ID starting with 58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52 not found: ID does not exist" containerID="58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.833990 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52"} err="failed to get container status \"58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52\": rpc error: code = NotFound desc = could not find container \"58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52\": container with ID starting with 58f0b4bfb7945931adf3f6a2160bb7a56ddc5ae00026ebcc245472d401bd7c52 not found: ID does not exist" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.834018 4696 scope.go:117] "RemoveContainer" containerID="be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c" Nov 25 11:10:05 crc kubenswrapper[4696]: E1125 11:10:05.834507 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c\": container with ID starting with be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c not found: ID does not exist" containerID="be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c" Nov 25 11:10:05 crc kubenswrapper[4696]: I1125 11:10:05.834551 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c"} err="failed to get container status \"be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c\": rpc error: code = NotFound desc = could not find container \"be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c\": container with ID starting with be16b1c18d1f26ecaf7ed85bcc821ee7ea0c2578f5cad60b67d87660062faf2c not found: ID does not exist" Nov 25 11:10:06 crc kubenswrapper[4696]: I1125 11:10:06.052485 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" path="/var/lib/kubelet/pods/b199b23f-df6a-409f-a6fc-fe60ecb4bb2f/volumes" Nov 25 11:10:06 crc kubenswrapper[4696]: I1125 11:10:06.464645 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:10:06 crc kubenswrapper[4696]: I1125 11:10:06.464725 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:10:06 crc kubenswrapper[4696]: I1125 11:10:06.513192 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:10:10 crc kubenswrapper[4696]: I1125 11:10:10.041984 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:10:10 crc kubenswrapper[4696]: I1125 11:10:10.786183 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"f1e679762b5e7ea7648b96640824c202d277b5a114be58c963e8229384fbf19e"} Nov 25 11:10:15 crc kubenswrapper[4696]: I1125 11:10:15.831072 4696 generic.go:334] "Generic (PLEG): container finished" podID="d69df5c8-e28a-4b73-9d2b-97087aaed351" containerID="a9f90fda84ace6e3c775b1c7dbb71a1ce436f10bcc2a3202ee342403f514fa36" exitCode=0 Nov 25 11:10:15 crc kubenswrapper[4696]: I1125 11:10:15.831316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" event={"ID":"d69df5c8-e28a-4b73-9d2b-97087aaed351","Type":"ContainerDied","Data":"a9f90fda84ace6e3c775b1c7dbb71a1ce436f10bcc2a3202ee342403f514fa36"} Nov 25 11:10:16 crc kubenswrapper[4696]: I1125 11:10:16.508885 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:10:16 crc kubenswrapper[4696]: I1125 11:10:16.554073 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pt2hf"] Nov 25 11:10:16 crc kubenswrapper[4696]: I1125 11:10:16.838418 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pt2hf" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="registry-server" containerID="cri-o://2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590" gracePeriod=2 Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.319354 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.328771 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.502320 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-inventory\") pod \"d69df5c8-e28a-4b73-9d2b-97087aaed351\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.502478 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-catalog-content\") pod \"bba18af9-27fa-4993-99a7-d397563e0aa5\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.502566 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7g92\" (UniqueName: \"kubernetes.io/projected/d69df5c8-e28a-4b73-9d2b-97087aaed351-kube-api-access-r7g92\") pod \"d69df5c8-e28a-4b73-9d2b-97087aaed351\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.502612 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhcm6\" (UniqueName: \"kubernetes.io/projected/bba18af9-27fa-4993-99a7-d397563e0aa5-kube-api-access-fhcm6\") pod \"bba18af9-27fa-4993-99a7-d397563e0aa5\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.502782 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-utilities\") pod \"bba18af9-27fa-4993-99a7-d397563e0aa5\" (UID: \"bba18af9-27fa-4993-99a7-d397563e0aa5\") " Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.502813 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-ssh-key\") pod \"d69df5c8-e28a-4b73-9d2b-97087aaed351\" (UID: \"d69df5c8-e28a-4b73-9d2b-97087aaed351\") " Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.504242 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-utilities" (OuterVolumeSpecName: "utilities") pod "bba18af9-27fa-4993-99a7-d397563e0aa5" (UID: "bba18af9-27fa-4993-99a7-d397563e0aa5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.508528 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d69df5c8-e28a-4b73-9d2b-97087aaed351-kube-api-access-r7g92" (OuterVolumeSpecName: "kube-api-access-r7g92") pod "d69df5c8-e28a-4b73-9d2b-97087aaed351" (UID: "d69df5c8-e28a-4b73-9d2b-97087aaed351"). InnerVolumeSpecName "kube-api-access-r7g92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.508958 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bba18af9-27fa-4993-99a7-d397563e0aa5-kube-api-access-fhcm6" (OuterVolumeSpecName: "kube-api-access-fhcm6") pod "bba18af9-27fa-4993-99a7-d397563e0aa5" (UID: "bba18af9-27fa-4993-99a7-d397563e0aa5"). InnerVolumeSpecName "kube-api-access-fhcm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.535944 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-inventory" (OuterVolumeSpecName: "inventory") pod "d69df5c8-e28a-4b73-9d2b-97087aaed351" (UID: "d69df5c8-e28a-4b73-9d2b-97087aaed351"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.536976 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d69df5c8-e28a-4b73-9d2b-97087aaed351" (UID: "d69df5c8-e28a-4b73-9d2b-97087aaed351"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.563303 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bba18af9-27fa-4993-99a7-d397563e0aa5" (UID: "bba18af9-27fa-4993-99a7-d397563e0aa5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.606216 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.606449 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.606569 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7g92\" (UniqueName: \"kubernetes.io/projected/d69df5c8-e28a-4b73-9d2b-97087aaed351-kube-api-access-r7g92\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.606641 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhcm6\" (UniqueName: \"kubernetes.io/projected/bba18af9-27fa-4993-99a7-d397563e0aa5-kube-api-access-fhcm6\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.606728 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba18af9-27fa-4993-99a7-d397563e0aa5-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.606810 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d69df5c8-e28a-4b73-9d2b-97087aaed351-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.850194 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.851391 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kg6v5" event={"ID":"d69df5c8-e28a-4b73-9d2b-97087aaed351","Type":"ContainerDied","Data":"78927f7da77d3598b3d27500f265bcc6e645553354d1efc15e0b97a5941ea875"} Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.851439 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78927f7da77d3598b3d27500f265bcc6e645553354d1efc15e0b97a5941ea875" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.853905 4696 generic.go:334] "Generic (PLEG): container finished" podID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerID="2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590" exitCode=0 Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.853943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerDied","Data":"2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590"} Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.853980 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pt2hf" event={"ID":"bba18af9-27fa-4993-99a7-d397563e0aa5","Type":"ContainerDied","Data":"6c5d6a6f265731775c4748a0f210ba595d9bb1e68e02eac9f7d11e9b2f64a01e"} Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.854001 4696 scope.go:117] "RemoveContainer" containerID="2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.854170 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pt2hf" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.892772 4696 scope.go:117] "RemoveContainer" containerID="4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.916579 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pt2hf"] Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.916740 4696 scope.go:117] "RemoveContainer" containerID="7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.932035 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pt2hf"] Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.952411 4696 scope.go:117] "RemoveContainer" containerID="2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.955324 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590\": container with ID starting with 2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590 not found: ID does not exist" containerID="2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.955367 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590"} err="failed to get container status \"2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590\": rpc error: code = NotFound desc = could not find container \"2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590\": container with ID starting with 2d734dfcb7c0436f149995b894d32ed50508ea283834d5c883feadeb5f0fa590 not found: ID does not exist" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.955398 4696 scope.go:117] "RemoveContainer" containerID="4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.960193 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495\": container with ID starting with 4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495 not found: ID does not exist" containerID="4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.960236 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495"} err="failed to get container status \"4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495\": rpc error: code = NotFound desc = could not find container \"4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495\": container with ID starting with 4dc89c0dbb3d2bee91367faaf74c06841a4af4c64947fce8394cb11570ce1495 not found: ID does not exist" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.960264 4696 scope.go:117] "RemoveContainer" containerID="7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.963066 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306\": container with ID starting with 7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306 not found: ID does not exist" containerID="7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.963250 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306"} err="failed to get container status \"7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306\": rpc error: code = NotFound desc = could not find container \"7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306\": container with ID starting with 7c0c10ee0c111953f2e84f55c7e8b265b609415a797c16fccb89e5135211b306 not found: ID does not exist" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.978803 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww"] Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979373 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="extract-utilities" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979398 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="extract-utilities" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979425 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="extract-utilities" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979433 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="extract-utilities" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979455 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="extract-content" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979464 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="extract-content" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979483 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="registry-server" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979491 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="registry-server" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979506 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="registry-server" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979514 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="registry-server" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979537 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69df5c8-e28a-4b73-9d2b-97087aaed351" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979546 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69df5c8-e28a-4b73-9d2b-97087aaed351" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:10:17 crc kubenswrapper[4696]: E1125 11:10:17.979561 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="extract-content" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979569 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="extract-content" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.979975 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b199b23f-df6a-409f-a6fc-fe60ecb4bb2f" containerName="registry-server" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.980003 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" containerName="registry-server" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.980029 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d69df5c8-e28a-4b73-9d2b-97087aaed351" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.980907 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.986216 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.986739 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.986968 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:10:17 crc kubenswrapper[4696]: I1125 11:10:17.987169 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.000449 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww"] Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.055063 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bba18af9-27fa-4993-99a7-d397563e0aa5" path="/var/lib/kubelet/pods/bba18af9-27fa-4993-99a7-d397563e0aa5/volumes" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.117511 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.117606 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.117663 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf2bl\" (UniqueName: \"kubernetes.io/projected/114399b2-d253-44d9-994e-5098ef72a972-kube-api-access-mf2bl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.220760 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.220989 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.221120 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf2bl\" (UniqueName: \"kubernetes.io/projected/114399b2-d253-44d9-994e-5098ef72a972-kube-api-access-mf2bl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.226962 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.227028 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.239127 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf2bl\" (UniqueName: \"kubernetes.io/projected/114399b2-d253-44d9-994e-5098ef72a972-kube-api-access-mf2bl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bzgww\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.331066 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:10:18 crc kubenswrapper[4696]: W1125 11:10:18.869505 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod114399b2_d253_44d9_994e_5098ef72a972.slice/crio-e301fa3b07daa4effdcd3eb02adfcebf429de2bbedca00334dbabb4182bf394c WatchSource:0}: Error finding container e301fa3b07daa4effdcd3eb02adfcebf429de2bbedca00334dbabb4182bf394c: Status 404 returned error can't find the container with id e301fa3b07daa4effdcd3eb02adfcebf429de2bbedca00334dbabb4182bf394c Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.871528 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww"] Nov 25 11:10:18 crc kubenswrapper[4696]: I1125 11:10:18.872176 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:10:19 crc kubenswrapper[4696]: I1125 11:10:19.878014 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" event={"ID":"114399b2-d253-44d9-994e-5098ef72a972","Type":"ContainerStarted","Data":"b1caf30664ab41e823180776014523aa8244bd95d880931ec3a15bc8af9a31bd"} Nov 25 11:10:19 crc kubenswrapper[4696]: I1125 11:10:19.878364 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" event={"ID":"114399b2-d253-44d9-994e-5098ef72a972","Type":"ContainerStarted","Data":"e301fa3b07daa4effdcd3eb02adfcebf429de2bbedca00334dbabb4182bf394c"} Nov 25 11:10:19 crc kubenswrapper[4696]: I1125 11:10:19.906252 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" podStartSLOduration=2.473599532 podStartE2EDuration="2.906231253s" podCreationTimestamp="2025-11-25 11:10:17 +0000 UTC" firstStartedPulling="2025-11-25 11:10:18.871953047 +0000 UTC m=+2015.685569634" lastFinishedPulling="2025-11-25 11:10:19.304584768 +0000 UTC m=+2016.118201355" observedRunningTime="2025-11-25 11:10:19.897712731 +0000 UTC m=+2016.711329319" watchObservedRunningTime="2025-11-25 11:10:19.906231253 +0000 UTC m=+2016.719847850" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.441908 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gzz8h"] Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.444431 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.466464 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gzz8h"] Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.509311 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-catalog-content\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.509615 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvjpl\" (UniqueName: \"kubernetes.io/projected/53df25bc-d173-4497-a163-899d0f1a716b-kube-api-access-xvjpl\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.509826 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-utilities\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.611495 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-utilities\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.611578 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-catalog-content\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.611724 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvjpl\" (UniqueName: \"kubernetes.io/projected/53df25bc-d173-4497-a163-899d0f1a716b-kube-api-access-xvjpl\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.612290 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-catalog-content\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.612323 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-utilities\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.646423 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvjpl\" (UniqueName: \"kubernetes.io/projected/53df25bc-d173-4497-a163-899d0f1a716b-kube-api-access-xvjpl\") pod \"redhat-operators-gzz8h\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:06 crc kubenswrapper[4696]: I1125 11:11:06.763580 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:07 crc kubenswrapper[4696]: I1125 11:11:07.567319 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gzz8h"] Nov 25 11:11:08 crc kubenswrapper[4696]: I1125 11:11:08.297599 4696 generic.go:334] "Generic (PLEG): container finished" podID="53df25bc-d173-4497-a163-899d0f1a716b" containerID="3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a" exitCode=0 Nov 25 11:11:08 crc kubenswrapper[4696]: I1125 11:11:08.298109 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerDied","Data":"3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a"} Nov 25 11:11:08 crc kubenswrapper[4696]: I1125 11:11:08.298194 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerStarted","Data":"0ac5e820b15f33973a3fe558b1e61a176603c22e711239fcc144f69ff95ac92b"} Nov 25 11:11:09 crc kubenswrapper[4696]: I1125 11:11:09.309984 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerStarted","Data":"e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7"} Nov 25 11:11:14 crc kubenswrapper[4696]: I1125 11:11:14.357588 4696 generic.go:334] "Generic (PLEG): container finished" podID="53df25bc-d173-4497-a163-899d0f1a716b" containerID="e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7" exitCode=0 Nov 25 11:11:14 crc kubenswrapper[4696]: I1125 11:11:14.357696 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerDied","Data":"e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7"} Nov 25 11:11:15 crc kubenswrapper[4696]: I1125 11:11:15.366281 4696 generic.go:334] "Generic (PLEG): container finished" podID="114399b2-d253-44d9-994e-5098ef72a972" containerID="b1caf30664ab41e823180776014523aa8244bd95d880931ec3a15bc8af9a31bd" exitCode=0 Nov 25 11:11:15 crc kubenswrapper[4696]: I1125 11:11:15.366625 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" event={"ID":"114399b2-d253-44d9-994e-5098ef72a972","Type":"ContainerDied","Data":"b1caf30664ab41e823180776014523aa8244bd95d880931ec3a15bc8af9a31bd"} Nov 25 11:11:15 crc kubenswrapper[4696]: I1125 11:11:15.369004 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerStarted","Data":"322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b"} Nov 25 11:11:15 crc kubenswrapper[4696]: I1125 11:11:15.436761 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gzz8h" podStartSLOduration=3.010978633 podStartE2EDuration="9.436744003s" podCreationTimestamp="2025-11-25 11:11:06 +0000 UTC" firstStartedPulling="2025-11-25 11:11:08.300143133 +0000 UTC m=+2065.113759720" lastFinishedPulling="2025-11-25 11:11:14.725908503 +0000 UTC m=+2071.539525090" observedRunningTime="2025-11-25 11:11:15.430445692 +0000 UTC m=+2072.244062289" watchObservedRunningTime="2025-11-25 11:11:15.436744003 +0000 UTC m=+2072.250360590" Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.763926 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.764280 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.818913 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.910079 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf2bl\" (UniqueName: \"kubernetes.io/projected/114399b2-d253-44d9-994e-5098ef72a972-kube-api-access-mf2bl\") pod \"114399b2-d253-44d9-994e-5098ef72a972\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.910627 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-inventory\") pod \"114399b2-d253-44d9-994e-5098ef72a972\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.910843 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-ssh-key\") pod \"114399b2-d253-44d9-994e-5098ef72a972\" (UID: \"114399b2-d253-44d9-994e-5098ef72a972\") " Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.928473 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/114399b2-d253-44d9-994e-5098ef72a972-kube-api-access-mf2bl" (OuterVolumeSpecName: "kube-api-access-mf2bl") pod "114399b2-d253-44d9-994e-5098ef72a972" (UID: "114399b2-d253-44d9-994e-5098ef72a972"). InnerVolumeSpecName "kube-api-access-mf2bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.955317 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "114399b2-d253-44d9-994e-5098ef72a972" (UID: "114399b2-d253-44d9-994e-5098ef72a972"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:16 crc kubenswrapper[4696]: I1125 11:11:16.958323 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-inventory" (OuterVolumeSpecName: "inventory") pod "114399b2-d253-44d9-994e-5098ef72a972" (UID: "114399b2-d253-44d9-994e-5098ef72a972"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.012752 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf2bl\" (UniqueName: \"kubernetes.io/projected/114399b2-d253-44d9-994e-5098ef72a972-kube-api-access-mf2bl\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.012784 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.012793 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/114399b2-d253-44d9-994e-5098ef72a972-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.385927 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" event={"ID":"114399b2-d253-44d9-994e-5098ef72a972","Type":"ContainerDied","Data":"e301fa3b07daa4effdcd3eb02adfcebf429de2bbedca00334dbabb4182bf394c"} Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.386262 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e301fa3b07daa4effdcd3eb02adfcebf429de2bbedca00334dbabb4182bf394c" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.386329 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bzgww" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.496619 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9w2md"] Nov 25 11:11:17 crc kubenswrapper[4696]: E1125 11:11:17.497266 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="114399b2-d253-44d9-994e-5098ef72a972" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.497336 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="114399b2-d253-44d9-994e-5098ef72a972" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.497584 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="114399b2-d253-44d9-994e-5098ef72a972" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.498295 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.502010 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.502255 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.502463 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.502678 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.511441 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9w2md"] Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.622566 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txw2q\" (UniqueName: \"kubernetes.io/projected/d97e4985-25b9-4106-8053-b7ee1aa42542-kube-api-access-txw2q\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.622908 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.623067 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.724731 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.724837 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.724932 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txw2q\" (UniqueName: \"kubernetes.io/projected/d97e4985-25b9-4106-8053-b7ee1aa42542-kube-api-access-txw2q\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.733082 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.734381 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.754400 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txw2q\" (UniqueName: \"kubernetes.io/projected/d97e4985-25b9-4106-8053-b7ee1aa42542-kube-api-access-txw2q\") pod \"ssh-known-hosts-edpm-deployment-9w2md\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.817136 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gzz8h" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:11:17 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:11:17 crc kubenswrapper[4696]: > Nov 25 11:11:17 crc kubenswrapper[4696]: I1125 11:11:17.822581 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:18 crc kubenswrapper[4696]: I1125 11:11:18.380625 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9w2md"] Nov 25 11:11:19 crc kubenswrapper[4696]: I1125 11:11:19.447247 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" event={"ID":"d97e4985-25b9-4106-8053-b7ee1aa42542","Type":"ContainerStarted","Data":"f1bbc793d6cd32f207fd83ab4d4c989675ddd2a6be00e699d4e1803f6559d3d4"} Nov 25 11:11:19 crc kubenswrapper[4696]: I1125 11:11:19.447779 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" event={"ID":"d97e4985-25b9-4106-8053-b7ee1aa42542","Type":"ContainerStarted","Data":"722beb6bd93838d619e8c9c78ae12dcd0000d980a02ed7348e0d436951823db8"} Nov 25 11:11:19 crc kubenswrapper[4696]: I1125 11:11:19.509589 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" podStartSLOduration=1.783505839 podStartE2EDuration="2.509567067s" podCreationTimestamp="2025-11-25 11:11:17 +0000 UTC" firstStartedPulling="2025-11-25 11:11:18.39711737 +0000 UTC m=+2075.210733967" lastFinishedPulling="2025-11-25 11:11:19.123178598 +0000 UTC m=+2075.936795195" observedRunningTime="2025-11-25 11:11:19.508160332 +0000 UTC m=+2076.321776919" watchObservedRunningTime="2025-11-25 11:11:19.509567067 +0000 UTC m=+2076.323183654" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.584214 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lzrtk"] Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.588187 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.606193 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzrtk"] Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.763091 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-catalog-content\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.763328 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfbwm\" (UniqueName: \"kubernetes.io/projected/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-kube-api-access-tfbwm\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.763363 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-utilities\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.865570 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfbwm\" (UniqueName: \"kubernetes.io/projected/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-kube-api-access-tfbwm\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.865631 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-utilities\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.865711 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-catalog-content\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.866303 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-catalog-content\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.866303 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-utilities\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.895412 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfbwm\" (UniqueName: \"kubernetes.io/projected/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-kube-api-access-tfbwm\") pod \"certified-operators-lzrtk\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:23 crc kubenswrapper[4696]: I1125 11:11:23.944344 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:24 crc kubenswrapper[4696]: I1125 11:11:24.430832 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzrtk"] Nov 25 11:11:24 crc kubenswrapper[4696]: I1125 11:11:24.523819 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerStarted","Data":"f682609785d4910c5136a416427bba3acb4f5470b10d8056069a957a08432510"} Nov 25 11:11:25 crc kubenswrapper[4696]: I1125 11:11:25.533748 4696 generic.go:334] "Generic (PLEG): container finished" podID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerID="83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc" exitCode=0 Nov 25 11:11:25 crc kubenswrapper[4696]: I1125 11:11:25.533863 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerDied","Data":"83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc"} Nov 25 11:11:27 crc kubenswrapper[4696]: I1125 11:11:27.554877 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerStarted","Data":"41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e"} Nov 25 11:11:27 crc kubenswrapper[4696]: I1125 11:11:27.557803 4696 generic.go:334] "Generic (PLEG): container finished" podID="d97e4985-25b9-4106-8053-b7ee1aa42542" containerID="f1bbc793d6cd32f207fd83ab4d4c989675ddd2a6be00e699d4e1803f6559d3d4" exitCode=0 Nov 25 11:11:27 crc kubenswrapper[4696]: I1125 11:11:27.557847 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" event={"ID":"d97e4985-25b9-4106-8053-b7ee1aa42542","Type":"ContainerDied","Data":"f1bbc793d6cd32f207fd83ab4d4c989675ddd2a6be00e699d4e1803f6559d3d4"} Nov 25 11:11:27 crc kubenswrapper[4696]: I1125 11:11:27.810790 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gzz8h" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="registry-server" probeResult="failure" output=< Nov 25 11:11:27 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:11:27 crc kubenswrapper[4696]: > Nov 25 11:11:28 crc kubenswrapper[4696]: I1125 11:11:28.570210 4696 generic.go:334] "Generic (PLEG): container finished" podID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerID="41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e" exitCode=0 Nov 25 11:11:28 crc kubenswrapper[4696]: I1125 11:11:28.570879 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerDied","Data":"41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e"} Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.246921 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.376582 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txw2q\" (UniqueName: \"kubernetes.io/projected/d97e4985-25b9-4106-8053-b7ee1aa42542-kube-api-access-txw2q\") pod \"d97e4985-25b9-4106-8053-b7ee1aa42542\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.376641 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-ssh-key-openstack-edpm-ipam\") pod \"d97e4985-25b9-4106-8053-b7ee1aa42542\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.376800 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-inventory-0\") pod \"d97e4985-25b9-4106-8053-b7ee1aa42542\" (UID: \"d97e4985-25b9-4106-8053-b7ee1aa42542\") " Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.394959 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97e4985-25b9-4106-8053-b7ee1aa42542-kube-api-access-txw2q" (OuterVolumeSpecName: "kube-api-access-txw2q") pod "d97e4985-25b9-4106-8053-b7ee1aa42542" (UID: "d97e4985-25b9-4106-8053-b7ee1aa42542"). InnerVolumeSpecName "kube-api-access-txw2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.407854 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d97e4985-25b9-4106-8053-b7ee1aa42542" (UID: "d97e4985-25b9-4106-8053-b7ee1aa42542"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.418923 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d97e4985-25b9-4106-8053-b7ee1aa42542" (UID: "d97e4985-25b9-4106-8053-b7ee1aa42542"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.482032 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txw2q\" (UniqueName: \"kubernetes.io/projected/d97e4985-25b9-4106-8053-b7ee1aa42542-kube-api-access-txw2q\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.482070 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.482080 4696 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d97e4985-25b9-4106-8053-b7ee1aa42542-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.580077 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" event={"ID":"d97e4985-25b9-4106-8053-b7ee1aa42542","Type":"ContainerDied","Data":"722beb6bd93838d619e8c9c78ae12dcd0000d980a02ed7348e0d436951823db8"} Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.580147 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="722beb6bd93838d619e8c9c78ae12dcd0000d980a02ed7348e0d436951823db8" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.580089 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9w2md" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.691611 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq"] Nov 25 11:11:29 crc kubenswrapper[4696]: E1125 11:11:29.692247 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97e4985-25b9-4106-8053-b7ee1aa42542" containerName="ssh-known-hosts-edpm-deployment" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.692265 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97e4985-25b9-4106-8053-b7ee1aa42542" containerName="ssh-known-hosts-edpm-deployment" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.692450 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d97e4985-25b9-4106-8053-b7ee1aa42542" containerName="ssh-known-hosts-edpm-deployment" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.693120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.699175 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.699487 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.699789 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.700358 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.702384 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq"] Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.889702 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.889824 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxztf\" (UniqueName: \"kubernetes.io/projected/438e2932-b4ff-4135-b4a8-7b4e89d574fc-kube-api-access-vxztf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.889854 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.991318 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.991741 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxztf\" (UniqueName: \"kubernetes.io/projected/438e2932-b4ff-4135-b4a8-7b4e89d574fc-kube-api-access-vxztf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.991783 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.995558 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:29 crc kubenswrapper[4696]: I1125 11:11:29.995584 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:30 crc kubenswrapper[4696]: I1125 11:11:30.010311 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxztf\" (UniqueName: \"kubernetes.io/projected/438e2932-b4ff-4135-b4a8-7b4e89d574fc-kube-api-access-vxztf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w8rsq\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:30 crc kubenswrapper[4696]: I1125 11:11:30.308467 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:30 crc kubenswrapper[4696]: I1125 11:11:30.591436 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerStarted","Data":"3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c"} Nov 25 11:11:30 crc kubenswrapper[4696]: I1125 11:11:30.621729 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lzrtk" podStartSLOduration=3.410753397 podStartE2EDuration="7.621704494s" podCreationTimestamp="2025-11-25 11:11:23 +0000 UTC" firstStartedPulling="2025-11-25 11:11:25.536481129 +0000 UTC m=+2082.350097716" lastFinishedPulling="2025-11-25 11:11:29.747432226 +0000 UTC m=+2086.561048813" observedRunningTime="2025-11-25 11:11:30.613558727 +0000 UTC m=+2087.427175314" watchObservedRunningTime="2025-11-25 11:11:30.621704494 +0000 UTC m=+2087.435321091" Nov 25 11:11:30 crc kubenswrapper[4696]: I1125 11:11:30.870225 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq"] Nov 25 11:11:31 crc kubenswrapper[4696]: I1125 11:11:31.601144 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" event={"ID":"438e2932-b4ff-4135-b4a8-7b4e89d574fc","Type":"ContainerStarted","Data":"e799d073d3240dff0f8fb4e313e7ad8f3bf4f1b4a33e71540eff059844dd63bd"} Nov 25 11:11:31 crc kubenswrapper[4696]: I1125 11:11:31.601746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" event={"ID":"438e2932-b4ff-4135-b4a8-7b4e89d574fc","Type":"ContainerStarted","Data":"2fb201a6da9ee35db31994b6404791dc04788fe057d10e0703bb38debd2717d0"} Nov 25 11:11:31 crc kubenswrapper[4696]: I1125 11:11:31.617802 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" podStartSLOduration=2.22083118 podStartE2EDuration="2.617786731s" podCreationTimestamp="2025-11-25 11:11:29 +0000 UTC" firstStartedPulling="2025-11-25 11:11:30.891411573 +0000 UTC m=+2087.705028160" lastFinishedPulling="2025-11-25 11:11:31.288367114 +0000 UTC m=+2088.101983711" observedRunningTime="2025-11-25 11:11:31.613124644 +0000 UTC m=+2088.426741231" watchObservedRunningTime="2025-11-25 11:11:31.617786731 +0000 UTC m=+2088.431403318" Nov 25 11:11:33 crc kubenswrapper[4696]: I1125 11:11:33.945707 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:33 crc kubenswrapper[4696]: I1125 11:11:33.946250 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:34 crc kubenswrapper[4696]: I1125 11:11:34.001166 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:34 crc kubenswrapper[4696]: I1125 11:11:34.669984 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:34 crc kubenswrapper[4696]: I1125 11:11:34.713924 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzrtk"] Nov 25 11:11:36 crc kubenswrapper[4696]: I1125 11:11:36.643053 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lzrtk" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="registry-server" containerID="cri-o://3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c" gracePeriod=2 Nov 25 11:11:36 crc kubenswrapper[4696]: I1125 11:11:36.811827 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:36 crc kubenswrapper[4696]: I1125 11:11:36.882926 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.065716 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.120396 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfbwm\" (UniqueName: \"kubernetes.io/projected/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-kube-api-access-tfbwm\") pod \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.120486 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-utilities\") pod \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.120619 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-catalog-content\") pod \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\" (UID: \"7051ebf8-cd1a-43b5-8c4b-58bf7512515b\") " Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.121181 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-utilities" (OuterVolumeSpecName: "utilities") pod "7051ebf8-cd1a-43b5-8c4b-58bf7512515b" (UID: "7051ebf8-cd1a-43b5-8c4b-58bf7512515b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.127521 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-kube-api-access-tfbwm" (OuterVolumeSpecName: "kube-api-access-tfbwm") pod "7051ebf8-cd1a-43b5-8c4b-58bf7512515b" (UID: "7051ebf8-cd1a-43b5-8c4b-58bf7512515b"). InnerVolumeSpecName "kube-api-access-tfbwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.166863 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7051ebf8-cd1a-43b5-8c4b-58bf7512515b" (UID: "7051ebf8-cd1a-43b5-8c4b-58bf7512515b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.223296 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfbwm\" (UniqueName: \"kubernetes.io/projected/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-kube-api-access-tfbwm\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.223346 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.223358 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7051ebf8-cd1a-43b5-8c4b-58bf7512515b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.655085 4696 generic.go:334] "Generic (PLEG): container finished" podID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerID="3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c" exitCode=0 Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.655175 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzrtk" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.655183 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerDied","Data":"3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c"} Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.655239 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzrtk" event={"ID":"7051ebf8-cd1a-43b5-8c4b-58bf7512515b","Type":"ContainerDied","Data":"f682609785d4910c5136a416427bba3acb4f5470b10d8056069a957a08432510"} Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.655264 4696 scope.go:117] "RemoveContainer" containerID="3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.693703 4696 scope.go:117] "RemoveContainer" containerID="41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.700085 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzrtk"] Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.709147 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lzrtk"] Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.740357 4696 scope.go:117] "RemoveContainer" containerID="83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.844617 4696 scope.go:117] "RemoveContainer" containerID="3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c" Nov 25 11:11:37 crc kubenswrapper[4696]: E1125 11:11:37.845158 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c\": container with ID starting with 3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c not found: ID does not exist" containerID="3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.845194 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c"} err="failed to get container status \"3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c\": rpc error: code = NotFound desc = could not find container \"3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c\": container with ID starting with 3c1a7254d3f4c25f0da5ff03885de5e7b0cd949107c12adca05336ed5979719c not found: ID does not exist" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.845238 4696 scope.go:117] "RemoveContainer" containerID="41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e" Nov 25 11:11:37 crc kubenswrapper[4696]: E1125 11:11:37.845480 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e\": container with ID starting with 41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e not found: ID does not exist" containerID="41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.845516 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e"} err="failed to get container status \"41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e\": rpc error: code = NotFound desc = could not find container \"41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e\": container with ID starting with 41eb295562d3da701dbb2f381e33a828a316aa01542869747f53777f67e3a09e not found: ID does not exist" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.845534 4696 scope.go:117] "RemoveContainer" containerID="83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc" Nov 25 11:11:37 crc kubenswrapper[4696]: E1125 11:11:37.845778 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc\": container with ID starting with 83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc not found: ID does not exist" containerID="83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc" Nov 25 11:11:37 crc kubenswrapper[4696]: I1125 11:11:37.845799 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc"} err="failed to get container status \"83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc\": rpc error: code = NotFound desc = could not find container \"83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc\": container with ID starting with 83cb5d6c2a2b9f0758e4fbbf1be63413000db834f652cbc3c5d53f4269a2bfbc not found: ID does not exist" Nov 25 11:11:38 crc kubenswrapper[4696]: I1125 11:11:38.053292 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" path="/var/lib/kubelet/pods/7051ebf8-cd1a-43b5-8c4b-58bf7512515b/volumes" Nov 25 11:11:38 crc kubenswrapper[4696]: I1125 11:11:38.054270 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gzz8h"] Nov 25 11:11:38 crc kubenswrapper[4696]: I1125 11:11:38.666005 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gzz8h" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="registry-server" containerID="cri-o://322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b" gracePeriod=2 Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.118302 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.159267 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-catalog-content\") pod \"53df25bc-d173-4497-a163-899d0f1a716b\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.159480 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-utilities\") pod \"53df25bc-d173-4497-a163-899d0f1a716b\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.159584 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvjpl\" (UniqueName: \"kubernetes.io/projected/53df25bc-d173-4497-a163-899d0f1a716b-kube-api-access-xvjpl\") pod \"53df25bc-d173-4497-a163-899d0f1a716b\" (UID: \"53df25bc-d173-4497-a163-899d0f1a716b\") " Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.160120 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-utilities" (OuterVolumeSpecName: "utilities") pod "53df25bc-d173-4497-a163-899d0f1a716b" (UID: "53df25bc-d173-4497-a163-899d0f1a716b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.176879 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53df25bc-d173-4497-a163-899d0f1a716b-kube-api-access-xvjpl" (OuterVolumeSpecName: "kube-api-access-xvjpl") pod "53df25bc-d173-4497-a163-899d0f1a716b" (UID: "53df25bc-d173-4497-a163-899d0f1a716b"). InnerVolumeSpecName "kube-api-access-xvjpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.262170 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.262208 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvjpl\" (UniqueName: \"kubernetes.io/projected/53df25bc-d173-4497-a163-899d0f1a716b-kube-api-access-xvjpl\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.265480 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53df25bc-d173-4497-a163-899d0f1a716b" (UID: "53df25bc-d173-4497-a163-899d0f1a716b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.364217 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53df25bc-d173-4497-a163-899d0f1a716b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.676928 4696 generic.go:334] "Generic (PLEG): container finished" podID="53df25bc-d173-4497-a163-899d0f1a716b" containerID="322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b" exitCode=0 Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.676981 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzz8h" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.677000 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerDied","Data":"322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b"} Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.677316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzz8h" event={"ID":"53df25bc-d173-4497-a163-899d0f1a716b","Type":"ContainerDied","Data":"0ac5e820b15f33973a3fe558b1e61a176603c22e711239fcc144f69ff95ac92b"} Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.677338 4696 scope.go:117] "RemoveContainer" containerID="322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.705997 4696 scope.go:117] "RemoveContainer" containerID="e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.716640 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gzz8h"] Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.726816 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gzz8h"] Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.742030 4696 scope.go:117] "RemoveContainer" containerID="3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.779451 4696 scope.go:117] "RemoveContainer" containerID="322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b" Nov 25 11:11:39 crc kubenswrapper[4696]: E1125 11:11:39.779900 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b\": container with ID starting with 322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b not found: ID does not exist" containerID="322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.779937 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b"} err="failed to get container status \"322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b\": rpc error: code = NotFound desc = could not find container \"322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b\": container with ID starting with 322039c71355f70292c9e0bd84fa2d28960a7ba57f94c593fe3f0c1f9565468b not found: ID does not exist" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.779973 4696 scope.go:117] "RemoveContainer" containerID="e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7" Nov 25 11:11:39 crc kubenswrapper[4696]: E1125 11:11:39.780928 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7\": container with ID starting with e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7 not found: ID does not exist" containerID="e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.780981 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7"} err="failed to get container status \"e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7\": rpc error: code = NotFound desc = could not find container \"e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7\": container with ID starting with e489d48ff0b2a8768252c4ed275df4123a87731ea6293dddf1af4ffdee32a0c7 not found: ID does not exist" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.781033 4696 scope.go:117] "RemoveContainer" containerID="3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a" Nov 25 11:11:39 crc kubenswrapper[4696]: E1125 11:11:39.781574 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a\": container with ID starting with 3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a not found: ID does not exist" containerID="3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a" Nov 25 11:11:39 crc kubenswrapper[4696]: I1125 11:11:39.781608 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a"} err="failed to get container status \"3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a\": rpc error: code = NotFound desc = could not find container \"3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a\": container with ID starting with 3188df4ddca491ddba30ead19a696332b88f3b60e0bcdc1e261333db6ef1882a not found: ID does not exist" Nov 25 11:11:40 crc kubenswrapper[4696]: I1125 11:11:40.054179 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53df25bc-d173-4497-a163-899d0f1a716b" path="/var/lib/kubelet/pods/53df25bc-d173-4497-a163-899d0f1a716b/volumes" Nov 25 11:11:40 crc kubenswrapper[4696]: I1125 11:11:40.706615 4696 generic.go:334] "Generic (PLEG): container finished" podID="438e2932-b4ff-4135-b4a8-7b4e89d574fc" containerID="e799d073d3240dff0f8fb4e313e7ad8f3bf4f1b4a33e71540eff059844dd63bd" exitCode=0 Nov 25 11:11:40 crc kubenswrapper[4696]: I1125 11:11:40.706821 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" event={"ID":"438e2932-b4ff-4135-b4a8-7b4e89d574fc","Type":"ContainerDied","Data":"e799d073d3240dff0f8fb4e313e7ad8f3bf4f1b4a33e71540eff059844dd63bd"} Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.112223 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.221892 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-inventory\") pod \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.222061 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxztf\" (UniqueName: \"kubernetes.io/projected/438e2932-b4ff-4135-b4a8-7b4e89d574fc-kube-api-access-vxztf\") pod \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.223805 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-ssh-key\") pod \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\" (UID: \"438e2932-b4ff-4135-b4a8-7b4e89d574fc\") " Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.266022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/438e2932-b4ff-4135-b4a8-7b4e89d574fc-kube-api-access-vxztf" (OuterVolumeSpecName: "kube-api-access-vxztf") pod "438e2932-b4ff-4135-b4a8-7b4e89d574fc" (UID: "438e2932-b4ff-4135-b4a8-7b4e89d574fc"). InnerVolumeSpecName "kube-api-access-vxztf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.305893 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "438e2932-b4ff-4135-b4a8-7b4e89d574fc" (UID: "438e2932-b4ff-4135-b4a8-7b4e89d574fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.328197 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.328229 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxztf\" (UniqueName: \"kubernetes.io/projected/438e2932-b4ff-4135-b4a8-7b4e89d574fc-kube-api-access-vxztf\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.362822 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-inventory" (OuterVolumeSpecName: "inventory") pod "438e2932-b4ff-4135-b4a8-7b4e89d574fc" (UID: "438e2932-b4ff-4135-b4a8-7b4e89d574fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.430202 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/438e2932-b4ff-4135-b4a8-7b4e89d574fc-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.723592 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" event={"ID":"438e2932-b4ff-4135-b4a8-7b4e89d574fc","Type":"ContainerDied","Data":"2fb201a6da9ee35db31994b6404791dc04788fe057d10e0703bb38debd2717d0"} Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.723893 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb201a6da9ee35db31994b6404791dc04788fe057d10e0703bb38debd2717d0" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.723709 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w8rsq" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.885786 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7"] Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.886719 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="registry-server" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.888442 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="registry-server" Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.888528 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="extract-content" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.888581 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="extract-content" Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.888648 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="extract-utilities" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.888728 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="extract-utilities" Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.888812 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="extract-content" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.888870 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="extract-content" Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.888927 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="registry-server" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.888976 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="registry-server" Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.889040 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="extract-utilities" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.889094 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="extract-utilities" Nov 25 11:11:42 crc kubenswrapper[4696]: E1125 11:11:42.889172 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438e2932-b4ff-4135-b4a8-7b4e89d574fc" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.889229 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="438e2932-b4ff-4135-b4a8-7b4e89d574fc" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.889542 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="438e2932-b4ff-4135-b4a8-7b4e89d574fc" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.889604 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="53df25bc-d173-4497-a163-899d0f1a716b" containerName="registry-server" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.889687 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7051ebf8-cd1a-43b5-8c4b-58bf7512515b" containerName="registry-server" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.890432 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.894801 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.895082 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.895497 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.905448 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.908781 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7"] Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.940586 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48szf\" (UniqueName: \"kubernetes.io/projected/5779201b-7609-43c1-9cec-c3cc15302542-kube-api-access-48szf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.940655 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:42 crc kubenswrapper[4696]: I1125 11:11:42.940810 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.042277 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.042389 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48szf\" (UniqueName: \"kubernetes.io/projected/5779201b-7609-43c1-9cec-c3cc15302542-kube-api-access-48szf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.042443 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.047565 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.047695 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.059596 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48szf\" (UniqueName: \"kubernetes.io/projected/5779201b-7609-43c1-9cec-c3cc15302542-kube-api-access-48szf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.216770 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:43 crc kubenswrapper[4696]: I1125 11:11:43.763863 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7"] Nov 25 11:11:43 crc kubenswrapper[4696]: W1125 11:11:43.768335 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5779201b_7609_43c1_9cec_c3cc15302542.slice/crio-28d82cb56717119939272996f41a03912c0664300f7bff290531b121e1ce39dc WatchSource:0}: Error finding container 28d82cb56717119939272996f41a03912c0664300f7bff290531b121e1ce39dc: Status 404 returned error can't find the container with id 28d82cb56717119939272996f41a03912c0664300f7bff290531b121e1ce39dc Nov 25 11:11:44 crc kubenswrapper[4696]: I1125 11:11:44.194353 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:11:44 crc kubenswrapper[4696]: I1125 11:11:44.743120 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" event={"ID":"5779201b-7609-43c1-9cec-c3cc15302542","Type":"ContainerStarted","Data":"40892fdfab4dcd9a742d276068782f1a6421d82c4d86f142f968f4d6561c4de2"} Nov 25 11:11:44 crc kubenswrapper[4696]: I1125 11:11:44.743451 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" event={"ID":"5779201b-7609-43c1-9cec-c3cc15302542","Type":"ContainerStarted","Data":"28d82cb56717119939272996f41a03912c0664300f7bff290531b121e1ce39dc"} Nov 25 11:11:44 crc kubenswrapper[4696]: I1125 11:11:44.763786 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" podStartSLOduration=2.345998399 podStartE2EDuration="2.763762114s" podCreationTimestamp="2025-11-25 11:11:42 +0000 UTC" firstStartedPulling="2025-11-25 11:11:43.7719202 +0000 UTC m=+2100.585536787" lastFinishedPulling="2025-11-25 11:11:44.189683915 +0000 UTC m=+2101.003300502" observedRunningTime="2025-11-25 11:11:44.761369689 +0000 UTC m=+2101.574986276" watchObservedRunningTime="2025-11-25 11:11:44.763762114 +0000 UTC m=+2101.577378701" Nov 25 11:11:54 crc kubenswrapper[4696]: I1125 11:11:54.837625 4696 generic.go:334] "Generic (PLEG): container finished" podID="5779201b-7609-43c1-9cec-c3cc15302542" containerID="40892fdfab4dcd9a742d276068782f1a6421d82c4d86f142f968f4d6561c4de2" exitCode=0 Nov 25 11:11:54 crc kubenswrapper[4696]: I1125 11:11:54.837721 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" event={"ID":"5779201b-7609-43c1-9cec-c3cc15302542","Type":"ContainerDied","Data":"40892fdfab4dcd9a742d276068782f1a6421d82c4d86f142f968f4d6561c4de2"} Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.404475 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.508249 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-ssh-key\") pod \"5779201b-7609-43c1-9cec-c3cc15302542\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.508529 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-inventory\") pod \"5779201b-7609-43c1-9cec-c3cc15302542\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.508731 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48szf\" (UniqueName: \"kubernetes.io/projected/5779201b-7609-43c1-9cec-c3cc15302542-kube-api-access-48szf\") pod \"5779201b-7609-43c1-9cec-c3cc15302542\" (UID: \"5779201b-7609-43c1-9cec-c3cc15302542\") " Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.518972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5779201b-7609-43c1-9cec-c3cc15302542-kube-api-access-48szf" (OuterVolumeSpecName: "kube-api-access-48szf") pod "5779201b-7609-43c1-9cec-c3cc15302542" (UID: "5779201b-7609-43c1-9cec-c3cc15302542"). InnerVolumeSpecName "kube-api-access-48szf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.537437 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-inventory" (OuterVolumeSpecName: "inventory") pod "5779201b-7609-43c1-9cec-c3cc15302542" (UID: "5779201b-7609-43c1-9cec-c3cc15302542"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.544046 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5779201b-7609-43c1-9cec-c3cc15302542" (UID: "5779201b-7609-43c1-9cec-c3cc15302542"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.611470 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.611515 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5779201b-7609-43c1-9cec-c3cc15302542-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.611526 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48szf\" (UniqueName: \"kubernetes.io/projected/5779201b-7609-43c1-9cec-c3cc15302542-kube-api-access-48szf\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.858621 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" event={"ID":"5779201b-7609-43c1-9cec-c3cc15302542","Type":"ContainerDied","Data":"28d82cb56717119939272996f41a03912c0664300f7bff290531b121e1ce39dc"} Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.858697 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28d82cb56717119939272996f41a03912c0664300f7bff290531b121e1ce39dc" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.858729 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.959174 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6"] Nov 25 11:11:56 crc kubenswrapper[4696]: E1125 11:11:56.959561 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5779201b-7609-43c1-9cec-c3cc15302542" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.959578 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5779201b-7609-43c1-9cec-c3cc15302542" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.959800 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5779201b-7609-43c1-9cec-c3cc15302542" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.960476 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.972526 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.978373 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.980236 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.980441 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.980595 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.980756 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.980887 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.981056 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 25 11:11:56 crc kubenswrapper[4696]: I1125 11:11:56.997345 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6"] Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018474 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018519 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018539 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018561 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018593 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018619 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018645 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2fqx\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-kube-api-access-b2fqx\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018697 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018721 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018745 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018789 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018818 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018841 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.018864 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.120831 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.120957 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121003 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121052 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121121 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121162 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121211 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121262 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121301 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121333 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2fqx\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-kube-api-access-b2fqx\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121372 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121417 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.121498 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.134137 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.136025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.138487 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.140695 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.143557 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.143955 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.144592 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.145015 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.145171 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.145993 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.147251 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.150477 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.150611 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2fqx\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-kube-api-access-b2fqx\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.150783 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.276761 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:11:57 crc kubenswrapper[4696]: I1125 11:11:57.927808 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6"] Nov 25 11:11:58 crc kubenswrapper[4696]: I1125 11:11:58.878478 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" event={"ID":"3202ae74-b34c-41aa-b717-6313d9e71bf4","Type":"ContainerStarted","Data":"1316575594af36cb666b6c32a6b9f451d856499aaa7b2ec4e8caa37f74a3eb36"} Nov 25 11:11:58 crc kubenswrapper[4696]: I1125 11:11:58.879069 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" event={"ID":"3202ae74-b34c-41aa-b717-6313d9e71bf4","Type":"ContainerStarted","Data":"8eaa6b429184fe575ed9a900cafcd5c4ff99530cd3467ec84467c8b0b0e23e49"} Nov 25 11:12:30 crc kubenswrapper[4696]: I1125 11:12:30.802046 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:12:30 crc kubenswrapper[4696]: I1125 11:12:30.802728 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:12:39 crc kubenswrapper[4696]: I1125 11:12:39.259260 4696 generic.go:334] "Generic (PLEG): container finished" podID="3202ae74-b34c-41aa-b717-6313d9e71bf4" containerID="1316575594af36cb666b6c32a6b9f451d856499aaa7b2ec4e8caa37f74a3eb36" exitCode=0 Nov 25 11:12:39 crc kubenswrapper[4696]: I1125 11:12:39.259344 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" event={"ID":"3202ae74-b34c-41aa-b717-6313d9e71bf4","Type":"ContainerDied","Data":"1316575594af36cb666b6c32a6b9f451d856499aaa7b2ec4e8caa37f74a3eb36"} Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.670218 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.767850 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.767928 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-libvirt-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.767984 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2fqx\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-kube-api-access-b2fqx\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768006 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768062 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768128 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-neutron-metadata-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768170 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ssh-key\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768193 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-bootstrap-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768225 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768275 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-nova-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768300 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-repo-setup-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768364 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-inventory\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768424 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ovn-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.768450 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-telemetry-combined-ca-bundle\") pod \"3202ae74-b34c-41aa-b717-6313d9e71bf4\" (UID: \"3202ae74-b34c-41aa-b717-6313d9e71bf4\") " Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.776161 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.776161 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.776263 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.776885 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.777189 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.777345 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.777346 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-kube-api-access-b2fqx" (OuterVolumeSpecName: "kube-api-access-b2fqx") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "kube-api-access-b2fqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.777424 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.778132 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.778483 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.778935 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.779773 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.797806 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-inventory" (OuterVolumeSpecName: "inventory") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.804536 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3202ae74-b34c-41aa-b717-6313d9e71bf4" (UID: "3202ae74-b34c-41aa-b717-6313d9e71bf4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871295 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871425 4696 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871514 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871572 4696 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871690 4696 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871760 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871819 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871874 4696 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871937 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.871996 4696 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.872061 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2fqx\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-kube-api-access-b2fqx\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.872231 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.872332 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3202ae74-b34c-41aa-b717-6313d9e71bf4-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:40 crc kubenswrapper[4696]: I1125 11:12:40.872413 4696 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3202ae74-b34c-41aa-b717-6313d9e71bf4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.281451 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" event={"ID":"3202ae74-b34c-41aa-b717-6313d9e71bf4","Type":"ContainerDied","Data":"8eaa6b429184fe575ed9a900cafcd5c4ff99530cd3467ec84467c8b0b0e23e49"} Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.281490 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eaa6b429184fe575ed9a900cafcd5c4ff99530cd3467ec84467c8b0b0e23e49" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.281625 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.401350 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7"] Nov 25 11:12:41 crc kubenswrapper[4696]: E1125 11:12:41.401883 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3202ae74-b34c-41aa-b717-6313d9e71bf4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.401916 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3202ae74-b34c-41aa-b717-6313d9e71bf4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.402230 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3202ae74-b34c-41aa-b717-6313d9e71bf4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.403020 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.407005 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.407350 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.408924 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.409187 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.409354 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.414404 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7"] Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.483006 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c85js\" (UniqueName: \"kubernetes.io/projected/348a4b45-dfe7-449f-8b03-871394bfba46-kube-api-access-c85js\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.483109 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/348a4b45-dfe7-449f-8b03-871394bfba46-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.483156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.483181 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.483419 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.585319 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c85js\" (UniqueName: \"kubernetes.io/projected/348a4b45-dfe7-449f-8b03-871394bfba46-kube-api-access-c85js\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.585420 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/348a4b45-dfe7-449f-8b03-871394bfba46-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.585459 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.585491 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.585559 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.586698 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/348a4b45-dfe7-449f-8b03-871394bfba46-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.588986 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.589441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.592831 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.605745 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c85js\" (UniqueName: \"kubernetes.io/projected/348a4b45-dfe7-449f-8b03-871394bfba46-kube-api-access-c85js\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zf8h7\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:41 crc kubenswrapper[4696]: I1125 11:12:41.722250 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:12:42 crc kubenswrapper[4696]: I1125 11:12:42.225568 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7"] Nov 25 11:12:42 crc kubenswrapper[4696]: I1125 11:12:42.290632 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" event={"ID":"348a4b45-dfe7-449f-8b03-871394bfba46","Type":"ContainerStarted","Data":"30ba5e4470f9d557d4315273438179fff8c0cc96dab838a0fbec31cc8d1f416c"} Nov 25 11:12:43 crc kubenswrapper[4696]: I1125 11:12:43.300039 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" event={"ID":"348a4b45-dfe7-449f-8b03-871394bfba46","Type":"ContainerStarted","Data":"6413e880a90daf20876ed04d0d91cf9d4432d2045e71d234da110e19fb441664"} Nov 25 11:12:43 crc kubenswrapper[4696]: I1125 11:12:43.321074 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" podStartSLOduration=1.816592162 podStartE2EDuration="2.321051006s" podCreationTimestamp="2025-11-25 11:12:41 +0000 UTC" firstStartedPulling="2025-11-25 11:12:42.23068199 +0000 UTC m=+2159.044298587" lastFinishedPulling="2025-11-25 11:12:42.735140844 +0000 UTC m=+2159.548757431" observedRunningTime="2025-11-25 11:12:43.315278907 +0000 UTC m=+2160.128895504" watchObservedRunningTime="2025-11-25 11:12:43.321051006 +0000 UTC m=+2160.134667593" Nov 25 11:13:00 crc kubenswrapper[4696]: I1125 11:13:00.802631 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:13:00 crc kubenswrapper[4696]: I1125 11:13:00.803471 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:13:30 crc kubenswrapper[4696]: I1125 11:13:30.802018 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:13:30 crc kubenswrapper[4696]: I1125 11:13:30.802854 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:13:30 crc kubenswrapper[4696]: I1125 11:13:30.802936 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:13:30 crc kubenswrapper[4696]: I1125 11:13:30.804174 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1e679762b5e7ea7648b96640824c202d277b5a114be58c963e8229384fbf19e"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:13:30 crc kubenswrapper[4696]: I1125 11:13:30.804295 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://f1e679762b5e7ea7648b96640824c202d277b5a114be58c963e8229384fbf19e" gracePeriod=600 Nov 25 11:13:31 crc kubenswrapper[4696]: I1125 11:13:31.720652 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="f1e679762b5e7ea7648b96640824c202d277b5a114be58c963e8229384fbf19e" exitCode=0 Nov 25 11:13:31 crc kubenswrapper[4696]: I1125 11:13:31.720727 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"f1e679762b5e7ea7648b96640824c202d277b5a114be58c963e8229384fbf19e"} Nov 25 11:13:31 crc kubenswrapper[4696]: I1125 11:13:31.721136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09"} Nov 25 11:13:31 crc kubenswrapper[4696]: I1125 11:13:31.721157 4696 scope.go:117] "RemoveContainer" containerID="b8dd127a720913d1f15b22e0feb2a59636fdd7c3480ce52e71eb53db0265f8e9" Nov 25 11:13:49 crc kubenswrapper[4696]: I1125 11:13:49.881711 4696 generic.go:334] "Generic (PLEG): container finished" podID="348a4b45-dfe7-449f-8b03-871394bfba46" containerID="6413e880a90daf20876ed04d0d91cf9d4432d2045e71d234da110e19fb441664" exitCode=0 Nov 25 11:13:49 crc kubenswrapper[4696]: I1125 11:13:49.881814 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" event={"ID":"348a4b45-dfe7-449f-8b03-871394bfba46","Type":"ContainerDied","Data":"6413e880a90daf20876ed04d0d91cf9d4432d2045e71d234da110e19fb441664"} Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.275543 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.290433 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-inventory\") pod \"348a4b45-dfe7-449f-8b03-871394bfba46\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.290567 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c85js\" (UniqueName: \"kubernetes.io/projected/348a4b45-dfe7-449f-8b03-871394bfba46-kube-api-access-c85js\") pod \"348a4b45-dfe7-449f-8b03-871394bfba46\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.290606 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ovn-combined-ca-bundle\") pod \"348a4b45-dfe7-449f-8b03-871394bfba46\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.290774 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ssh-key\") pod \"348a4b45-dfe7-449f-8b03-871394bfba46\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.290892 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/348a4b45-dfe7-449f-8b03-871394bfba46-ovncontroller-config-0\") pod \"348a4b45-dfe7-449f-8b03-871394bfba46\" (UID: \"348a4b45-dfe7-449f-8b03-871394bfba46\") " Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.307831 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "348a4b45-dfe7-449f-8b03-871394bfba46" (UID: "348a4b45-dfe7-449f-8b03-871394bfba46"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.317174 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/348a4b45-dfe7-449f-8b03-871394bfba46-kube-api-access-c85js" (OuterVolumeSpecName: "kube-api-access-c85js") pod "348a4b45-dfe7-449f-8b03-871394bfba46" (UID: "348a4b45-dfe7-449f-8b03-871394bfba46"). InnerVolumeSpecName "kube-api-access-c85js". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.343570 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/348a4b45-dfe7-449f-8b03-871394bfba46-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "348a4b45-dfe7-449f-8b03-871394bfba46" (UID: "348a4b45-dfe7-449f-8b03-871394bfba46"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.346572 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "348a4b45-dfe7-449f-8b03-871394bfba46" (UID: "348a4b45-dfe7-449f-8b03-871394bfba46"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.352756 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-inventory" (OuterVolumeSpecName: "inventory") pod "348a4b45-dfe7-449f-8b03-871394bfba46" (UID: "348a4b45-dfe7-449f-8b03-871394bfba46"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.393192 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c85js\" (UniqueName: \"kubernetes.io/projected/348a4b45-dfe7-449f-8b03-871394bfba46-kube-api-access-c85js\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.393232 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.393241 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.393250 4696 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/348a4b45-dfe7-449f-8b03-871394bfba46-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.393259 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348a4b45-dfe7-449f-8b03-871394bfba46-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.898646 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" event={"ID":"348a4b45-dfe7-449f-8b03-871394bfba46","Type":"ContainerDied","Data":"30ba5e4470f9d557d4315273438179fff8c0cc96dab838a0fbec31cc8d1f416c"} Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.898709 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30ba5e4470f9d557d4315273438179fff8c0cc96dab838a0fbec31cc8d1f416c" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.898741 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zf8h7" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.999404 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns"] Nov 25 11:13:51 crc kubenswrapper[4696]: E1125 11:13:51.999825 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348a4b45-dfe7-449f-8b03-871394bfba46" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 11:13:51 crc kubenswrapper[4696]: I1125 11:13:51.999842 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="348a4b45-dfe7-449f-8b03-871394bfba46" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.000003 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="348a4b45-dfe7-449f-8b03-871394bfba46" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.000622 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.003458 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.003617 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.003818 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.003935 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.004070 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.009931 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.013877 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns"] Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.109581 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlwt7\" (UniqueName: \"kubernetes.io/projected/9813095c-910d-451b-b3bb-18a6d5fcdf26-kube-api-access-dlwt7\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.109888 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.110224 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.110326 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.110392 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.110443 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.212438 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlwt7\" (UniqueName: \"kubernetes.io/projected/9813095c-910d-451b-b3bb-18a6d5fcdf26-kube-api-access-dlwt7\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.212568 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.212636 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.212704 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.212741 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.212768 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.216557 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.216717 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.216853 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.217276 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.217613 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.232923 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlwt7\" (UniqueName: \"kubernetes.io/projected/9813095c-910d-451b-b3bb-18a6d5fcdf26-kube-api-access-dlwt7\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.318910 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.852613 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns"] Nov 25 11:13:52 crc kubenswrapper[4696]: I1125 11:13:52.929560 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" event={"ID":"9813095c-910d-451b-b3bb-18a6d5fcdf26","Type":"ContainerStarted","Data":"ca8a7d7c1064da94f800eb47ed5840004e959ac9da8eaad577a1945d5b4de11b"} Nov 25 11:13:53 crc kubenswrapper[4696]: I1125 11:13:53.939016 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" event={"ID":"9813095c-910d-451b-b3bb-18a6d5fcdf26","Type":"ContainerStarted","Data":"6acc0f4caac00b10eb95e80bc341f9be0d11a757e0a48c3470970b1180427701"} Nov 25 11:13:53 crc kubenswrapper[4696]: I1125 11:13:53.973437 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" podStartSLOduration=2.357275756 podStartE2EDuration="2.973415307s" podCreationTimestamp="2025-11-25 11:13:51 +0000 UTC" firstStartedPulling="2025-11-25 11:13:52.871986205 +0000 UTC m=+2229.685602802" lastFinishedPulling="2025-11-25 11:13:53.488125726 +0000 UTC m=+2230.301742353" observedRunningTime="2025-11-25 11:13:53.96433056 +0000 UTC m=+2230.777947177" watchObservedRunningTime="2025-11-25 11:13:53.973415307 +0000 UTC m=+2230.787031894" Nov 25 11:14:44 crc kubenswrapper[4696]: I1125 11:14:44.403800 4696 generic.go:334] "Generic (PLEG): container finished" podID="9813095c-910d-451b-b3bb-18a6d5fcdf26" containerID="6acc0f4caac00b10eb95e80bc341f9be0d11a757e0a48c3470970b1180427701" exitCode=0 Nov 25 11:14:44 crc kubenswrapper[4696]: I1125 11:14:44.404326 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" event={"ID":"9813095c-910d-451b-b3bb-18a6d5fcdf26","Type":"ContainerDied","Data":"6acc0f4caac00b10eb95e80bc341f9be0d11a757e0a48c3470970b1180427701"} Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.839745 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.914066 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-inventory\") pod \"9813095c-910d-451b-b3bb-18a6d5fcdf26\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.914509 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-nova-metadata-neutron-config-0\") pod \"9813095c-910d-451b-b3bb-18a6d5fcdf26\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.914603 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlwt7\" (UniqueName: \"kubernetes.io/projected/9813095c-910d-451b-b3bb-18a6d5fcdf26-kube-api-access-dlwt7\") pod \"9813095c-910d-451b-b3bb-18a6d5fcdf26\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.914696 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-metadata-combined-ca-bundle\") pod \"9813095c-910d-451b-b3bb-18a6d5fcdf26\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.914719 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-ssh-key\") pod \"9813095c-910d-451b-b3bb-18a6d5fcdf26\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.914738 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-ovn-metadata-agent-neutron-config-0\") pod \"9813095c-910d-451b-b3bb-18a6d5fcdf26\" (UID: \"9813095c-910d-451b-b3bb-18a6d5fcdf26\") " Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.921177 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9813095c-910d-451b-b3bb-18a6d5fcdf26" (UID: "9813095c-910d-451b-b3bb-18a6d5fcdf26"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.921571 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9813095c-910d-451b-b3bb-18a6d5fcdf26-kube-api-access-dlwt7" (OuterVolumeSpecName: "kube-api-access-dlwt7") pod "9813095c-910d-451b-b3bb-18a6d5fcdf26" (UID: "9813095c-910d-451b-b3bb-18a6d5fcdf26"). InnerVolumeSpecName "kube-api-access-dlwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.948307 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-inventory" (OuterVolumeSpecName: "inventory") pod "9813095c-910d-451b-b3bb-18a6d5fcdf26" (UID: "9813095c-910d-451b-b3bb-18a6d5fcdf26"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.950933 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9813095c-910d-451b-b3bb-18a6d5fcdf26" (UID: "9813095c-910d-451b-b3bb-18a6d5fcdf26"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.951321 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "9813095c-910d-451b-b3bb-18a6d5fcdf26" (UID: "9813095c-910d-451b-b3bb-18a6d5fcdf26"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:45 crc kubenswrapper[4696]: I1125 11:14:45.953406 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "9813095c-910d-451b-b3bb-18a6d5fcdf26" (UID: "9813095c-910d-451b-b3bb-18a6d5fcdf26"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.016846 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlwt7\" (UniqueName: \"kubernetes.io/projected/9813095c-910d-451b-b3bb-18a6d5fcdf26-kube-api-access-dlwt7\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.016896 4696 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.016915 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.016927 4696 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.016941 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.016950 4696 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9813095c-910d-451b-b3bb-18a6d5fcdf26-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.424600 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" event={"ID":"9813095c-910d-451b-b3bb-18a6d5fcdf26","Type":"ContainerDied","Data":"ca8a7d7c1064da94f800eb47ed5840004e959ac9da8eaad577a1945d5b4de11b"} Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.424653 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca8a7d7c1064da94f800eb47ed5840004e959ac9da8eaad577a1945d5b4de11b" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.424742 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.516213 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr"] Nov 25 11:14:46 crc kubenswrapper[4696]: E1125 11:14:46.516698 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9813095c-910d-451b-b3bb-18a6d5fcdf26" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.516721 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9813095c-910d-451b-b3bb-18a6d5fcdf26" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.516938 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9813095c-910d-451b-b3bb-18a6d5fcdf26" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.517597 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.520030 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.520275 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.520564 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.520807 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.527117 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.531982 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr"] Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.627606 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8ws\" (UniqueName: \"kubernetes.io/projected/27d6547e-1c82-4b39-a281-476463f32e81-kube-api-access-vw8ws\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.627700 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.627774 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.627820 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.627916 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.730113 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8ws\" (UniqueName: \"kubernetes.io/projected/27d6547e-1c82-4b39-a281-476463f32e81-kube-api-access-vw8ws\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.730234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.730308 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.730390 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.730598 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.735330 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.735763 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.736402 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.736988 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.749340 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8ws\" (UniqueName: \"kubernetes.io/projected/27d6547e-1c82-4b39-a281-476463f32e81-kube-api-access-vw8ws\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:46 crc kubenswrapper[4696]: I1125 11:14:46.838479 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:14:47 crc kubenswrapper[4696]: I1125 11:14:47.399014 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr"] Nov 25 11:14:47 crc kubenswrapper[4696]: I1125 11:14:47.436596 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" event={"ID":"27d6547e-1c82-4b39-a281-476463f32e81","Type":"ContainerStarted","Data":"6a8b905cb57f7862d3c1fff8906cb384794281c70de06cfc2e10dffe04c0504e"} Nov 25 11:14:48 crc kubenswrapper[4696]: I1125 11:14:48.447605 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" event={"ID":"27d6547e-1c82-4b39-a281-476463f32e81","Type":"ContainerStarted","Data":"7d7b6cb7506ce1c68ce2a29c423e36d9610b2ccb4d90f7cd82d918646de83c39"} Nov 25 11:14:48 crc kubenswrapper[4696]: I1125 11:14:48.476757 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" podStartSLOduration=2.028947347 podStartE2EDuration="2.476733934s" podCreationTimestamp="2025-11-25 11:14:46 +0000 UTC" firstStartedPulling="2025-11-25 11:14:47.406030082 +0000 UTC m=+2284.219646669" lastFinishedPulling="2025-11-25 11:14:47.853816629 +0000 UTC m=+2284.667433256" observedRunningTime="2025-11-25 11:14:48.467277476 +0000 UTC m=+2285.280894063" watchObservedRunningTime="2025-11-25 11:14:48.476733934 +0000 UTC m=+2285.290350521" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.136699 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd"] Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.140466 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.142233 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.143707 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.155911 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd"] Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.228125 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64ab406-d59c-4369-9ca9-91c47ee6887e-config-volume\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.228240 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwj46\" (UniqueName: \"kubernetes.io/projected/a64ab406-d59c-4369-9ca9-91c47ee6887e-kube-api-access-kwj46\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.228300 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64ab406-d59c-4369-9ca9-91c47ee6887e-secret-volume\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.330432 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64ab406-d59c-4369-9ca9-91c47ee6887e-secret-volume\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.330555 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64ab406-d59c-4369-9ca9-91c47ee6887e-config-volume\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.330682 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwj46\" (UniqueName: \"kubernetes.io/projected/a64ab406-d59c-4369-9ca9-91c47ee6887e-kube-api-access-kwj46\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.331860 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64ab406-d59c-4369-9ca9-91c47ee6887e-config-volume\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.336781 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64ab406-d59c-4369-9ca9-91c47ee6887e-secret-volume\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.347299 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwj46\" (UniqueName: \"kubernetes.io/projected/a64ab406-d59c-4369-9ca9-91c47ee6887e-kube-api-access-kwj46\") pod \"collect-profiles-29401155-z22qd\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.477857 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:00 crc kubenswrapper[4696]: I1125 11:15:00.924362 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd"] Nov 25 11:15:01 crc kubenswrapper[4696]: I1125 11:15:01.565287 4696 generic.go:334] "Generic (PLEG): container finished" podID="a64ab406-d59c-4369-9ca9-91c47ee6887e" containerID="6a58d1fa30495bdc32e09ac77dc6c1189b63e258dffe865ac90df83bf8ead2dd" exitCode=0 Nov 25 11:15:01 crc kubenswrapper[4696]: I1125 11:15:01.565330 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" event={"ID":"a64ab406-d59c-4369-9ca9-91c47ee6887e","Type":"ContainerDied","Data":"6a58d1fa30495bdc32e09ac77dc6c1189b63e258dffe865ac90df83bf8ead2dd"} Nov 25 11:15:01 crc kubenswrapper[4696]: I1125 11:15:01.565579 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" event={"ID":"a64ab406-d59c-4369-9ca9-91c47ee6887e","Type":"ContainerStarted","Data":"d7c43e825fc6de271c984dab4b47b65101de47a9a5dc43391dd87965d89a2fd5"} Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.901346 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.980215 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64ab406-d59c-4369-9ca9-91c47ee6887e-secret-volume\") pod \"a64ab406-d59c-4369-9ca9-91c47ee6887e\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.980374 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwj46\" (UniqueName: \"kubernetes.io/projected/a64ab406-d59c-4369-9ca9-91c47ee6887e-kube-api-access-kwj46\") pod \"a64ab406-d59c-4369-9ca9-91c47ee6887e\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.981017 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64ab406-d59c-4369-9ca9-91c47ee6887e-config-volume\") pod \"a64ab406-d59c-4369-9ca9-91c47ee6887e\" (UID: \"a64ab406-d59c-4369-9ca9-91c47ee6887e\") " Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.981584 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a64ab406-d59c-4369-9ca9-91c47ee6887e-config-volume" (OuterVolumeSpecName: "config-volume") pod "a64ab406-d59c-4369-9ca9-91c47ee6887e" (UID: "a64ab406-d59c-4369-9ca9-91c47ee6887e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.985966 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64ab406-d59c-4369-9ca9-91c47ee6887e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a64ab406-d59c-4369-9ca9-91c47ee6887e" (UID: "a64ab406-d59c-4369-9ca9-91c47ee6887e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:15:02 crc kubenswrapper[4696]: I1125 11:15:02.986061 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64ab406-d59c-4369-9ca9-91c47ee6887e-kube-api-access-kwj46" (OuterVolumeSpecName: "kube-api-access-kwj46") pod "a64ab406-d59c-4369-9ca9-91c47ee6887e" (UID: "a64ab406-d59c-4369-9ca9-91c47ee6887e"). InnerVolumeSpecName "kube-api-access-kwj46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.083074 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwj46\" (UniqueName: \"kubernetes.io/projected/a64ab406-d59c-4369-9ca9-91c47ee6887e-kube-api-access-kwj46\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.083113 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a64ab406-d59c-4369-9ca9-91c47ee6887e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.083122 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a64ab406-d59c-4369-9ca9-91c47ee6887e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.590376 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" event={"ID":"a64ab406-d59c-4369-9ca9-91c47ee6887e","Type":"ContainerDied","Data":"d7c43e825fc6de271c984dab4b47b65101de47a9a5dc43391dd87965d89a2fd5"} Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.590418 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c43e825fc6de271c984dab4b47b65101de47a9a5dc43391dd87965d89a2fd5" Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.590431 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd" Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.978761 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk"] Nov 25 11:15:03 crc kubenswrapper[4696]: I1125 11:15:03.987218 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-46chk"] Nov 25 11:15:04 crc kubenswrapper[4696]: I1125 11:15:04.076970 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6edd3c-d469-4a55-a9b8-2a3087248db6" path="/var/lib/kubelet/pods/0e6edd3c-d469-4a55-a9b8-2a3087248db6/volumes" Nov 25 11:15:36 crc kubenswrapper[4696]: I1125 11:15:36.485061 4696 scope.go:117] "RemoveContainer" containerID="9881f9e2a38942d186ed0835edac34e728e29c17093199894b1b465a1db89cc8" Nov 25 11:16:00 crc kubenswrapper[4696]: I1125 11:16:00.801783 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:16:00 crc kubenswrapper[4696]: I1125 11:16:00.802344 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:16:30 crc kubenswrapper[4696]: I1125 11:16:30.801459 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:16:30 crc kubenswrapper[4696]: I1125 11:16:30.801975 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:16:46 crc kubenswrapper[4696]: I1125 11:16:46.777202 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-bbdns" podUID="020fd783-0219-4f9d-8504-f430877d9ed4" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 11:17:00 crc kubenswrapper[4696]: I1125 11:17:00.803356 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:17:00 crc kubenswrapper[4696]: I1125 11:17:00.804013 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:17:00 crc kubenswrapper[4696]: I1125 11:17:00.804099 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:17:00 crc kubenswrapper[4696]: I1125 11:17:00.804945 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:17:00 crc kubenswrapper[4696]: I1125 11:17:00.805015 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" gracePeriod=600 Nov 25 11:17:00 crc kubenswrapper[4696]: E1125 11:17:00.929728 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:17:01 crc kubenswrapper[4696]: I1125 11:17:01.703828 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" exitCode=0 Nov 25 11:17:01 crc kubenswrapper[4696]: I1125 11:17:01.703922 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09"} Nov 25 11:17:01 crc kubenswrapper[4696]: I1125 11:17:01.704221 4696 scope.go:117] "RemoveContainer" containerID="f1e679762b5e7ea7648b96640824c202d277b5a114be58c963e8229384fbf19e" Nov 25 11:17:01 crc kubenswrapper[4696]: I1125 11:17:01.704935 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:17:01 crc kubenswrapper[4696]: E1125 11:17:01.705347 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:17:17 crc kubenswrapper[4696]: I1125 11:17:17.042488 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:17:17 crc kubenswrapper[4696]: E1125 11:17:17.044034 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:17:31 crc kubenswrapper[4696]: I1125 11:17:31.043211 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:17:31 crc kubenswrapper[4696]: E1125 11:17:31.045298 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:17:44 crc kubenswrapper[4696]: I1125 11:17:44.051209 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:17:44 crc kubenswrapper[4696]: E1125 11:17:44.052432 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:17:56 crc kubenswrapper[4696]: I1125 11:17:56.043048 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:17:56 crc kubenswrapper[4696]: E1125 11:17:56.043998 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:18:09 crc kubenswrapper[4696]: I1125 11:18:09.042940 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:18:09 crc kubenswrapper[4696]: E1125 11:18:09.043756 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:18:24 crc kubenswrapper[4696]: I1125 11:18:24.051776 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:18:24 crc kubenswrapper[4696]: E1125 11:18:24.052560 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:18:36 crc kubenswrapper[4696]: I1125 11:18:36.042932 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:18:36 crc kubenswrapper[4696]: E1125 11:18:36.044291 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:18:47 crc kubenswrapper[4696]: I1125 11:18:47.042987 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:18:47 crc kubenswrapper[4696]: E1125 11:18:47.044117 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:19:00 crc kubenswrapper[4696]: I1125 11:19:00.042330 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:19:00 crc kubenswrapper[4696]: E1125 11:19:00.043225 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:19:13 crc kubenswrapper[4696]: I1125 11:19:13.042734 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:19:13 crc kubenswrapper[4696]: E1125 11:19:13.043532 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:19:25 crc kubenswrapper[4696]: I1125 11:19:25.042431 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:19:25 crc kubenswrapper[4696]: E1125 11:19:25.043196 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:19:37 crc kubenswrapper[4696]: I1125 11:19:37.042164 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:19:37 crc kubenswrapper[4696]: E1125 11:19:37.042892 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:19:48 crc kubenswrapper[4696]: I1125 11:19:48.042553 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:19:48 crc kubenswrapper[4696]: E1125 11:19:48.043455 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:19:55 crc kubenswrapper[4696]: I1125 11:19:55.268982 4696 generic.go:334] "Generic (PLEG): container finished" podID="27d6547e-1c82-4b39-a281-476463f32e81" containerID="7d7b6cb7506ce1c68ce2a29c423e36d9610b2ccb4d90f7cd82d918646de83c39" exitCode=0 Nov 25 11:19:55 crc kubenswrapper[4696]: I1125 11:19:55.269059 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" event={"ID":"27d6547e-1c82-4b39-a281-476463f32e81","Type":"ContainerDied","Data":"7d7b6cb7506ce1c68ce2a29c423e36d9610b2ccb4d90f7cd82d918646de83c39"} Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.724506 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.755419 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-inventory\") pod \"27d6547e-1c82-4b39-a281-476463f32e81\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.755749 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-ssh-key\") pod \"27d6547e-1c82-4b39-a281-476463f32e81\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.755812 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-secret-0\") pod \"27d6547e-1c82-4b39-a281-476463f32e81\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.755850 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw8ws\" (UniqueName: \"kubernetes.io/projected/27d6547e-1c82-4b39-a281-476463f32e81-kube-api-access-vw8ws\") pod \"27d6547e-1c82-4b39-a281-476463f32e81\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.755909 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-combined-ca-bundle\") pod \"27d6547e-1c82-4b39-a281-476463f32e81\" (UID: \"27d6547e-1c82-4b39-a281-476463f32e81\") " Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.763752 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d6547e-1c82-4b39-a281-476463f32e81-kube-api-access-vw8ws" (OuterVolumeSpecName: "kube-api-access-vw8ws") pod "27d6547e-1c82-4b39-a281-476463f32e81" (UID: "27d6547e-1c82-4b39-a281-476463f32e81"). InnerVolumeSpecName "kube-api-access-vw8ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.790746 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "27d6547e-1c82-4b39-a281-476463f32e81" (UID: "27d6547e-1c82-4b39-a281-476463f32e81"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.791781 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-inventory" (OuterVolumeSpecName: "inventory") pod "27d6547e-1c82-4b39-a281-476463f32e81" (UID: "27d6547e-1c82-4b39-a281-476463f32e81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.804441 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "27d6547e-1c82-4b39-a281-476463f32e81" (UID: "27d6547e-1c82-4b39-a281-476463f32e81"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.821451 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27d6547e-1c82-4b39-a281-476463f32e81" (UID: "27d6547e-1c82-4b39-a281-476463f32e81"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.859082 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.859133 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.859143 4696 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.859155 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw8ws\" (UniqueName: \"kubernetes.io/projected/27d6547e-1c82-4b39-a281-476463f32e81-kube-api-access-vw8ws\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:56 crc kubenswrapper[4696]: I1125 11:19:56.859163 4696 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d6547e-1c82-4b39-a281-476463f32e81-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.317126 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" event={"ID":"27d6547e-1c82-4b39-a281-476463f32e81","Type":"ContainerDied","Data":"6a8b905cb57f7862d3c1fff8906cb384794281c70de06cfc2e10dffe04c0504e"} Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.317180 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a8b905cb57f7862d3c1fff8906cb384794281c70de06cfc2e10dffe04c0504e" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.317267 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.401428 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4"] Nov 25 11:19:57 crc kubenswrapper[4696]: E1125 11:19:57.401804 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64ab406-d59c-4369-9ca9-91c47ee6887e" containerName="collect-profiles" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.401820 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64ab406-d59c-4369-9ca9-91c47ee6887e" containerName="collect-profiles" Nov 25 11:19:57 crc kubenswrapper[4696]: E1125 11:19:57.401842 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d6547e-1c82-4b39-a281-476463f32e81" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.401851 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d6547e-1c82-4b39-a281-476463f32e81" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.402035 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64ab406-d59c-4369-9ca9-91c47ee6887e" containerName="collect-profiles" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.402067 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d6547e-1c82-4b39-a281-476463f32e81" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.402655 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.406436 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.406488 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.406740 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.406794 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.406873 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.407393 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.413440 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.418772 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4"] Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.470654 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.470751 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf6kp\" (UniqueName: \"kubernetes.io/projected/11902ad3-7187-4749-8fbf-3d5f1eb77764-kube-api-access-rf6kp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.470795 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.470822 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.470876 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.470901 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.471078 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.471138 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.471213 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573190 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573279 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573501 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf6kp\" (UniqueName: \"kubernetes.io/projected/11902ad3-7187-4749-8fbf-3d5f1eb77764-kube-api-access-rf6kp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573633 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573714 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573806 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.573852 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.574136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.574183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.574605 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.580827 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.580937 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.581249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.581419 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.581454 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.581570 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.581806 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.593970 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf6kp\" (UniqueName: \"kubernetes.io/projected/11902ad3-7187-4749-8fbf-3d5f1eb77764-kube-api-access-rf6kp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fkn4\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:57 crc kubenswrapper[4696]: I1125 11:19:57.724505 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:19:58 crc kubenswrapper[4696]: I1125 11:19:58.254998 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4"] Nov 25 11:19:58 crc kubenswrapper[4696]: I1125 11:19:58.265678 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:19:58 crc kubenswrapper[4696]: I1125 11:19:58.326501 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" event={"ID":"11902ad3-7187-4749-8fbf-3d5f1eb77764","Type":"ContainerStarted","Data":"fd427399fc2cdb7de2b1bda64803fc84dfd63c2ebb720c0065f08d246c5abb65"} Nov 25 11:19:59 crc kubenswrapper[4696]: I1125 11:19:59.336947 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" event={"ID":"11902ad3-7187-4749-8fbf-3d5f1eb77764","Type":"ContainerStarted","Data":"68259a8a34701b40f657503776589fd7374cc2fbf65415eaac39e1d514760dd2"} Nov 25 11:19:59 crc kubenswrapper[4696]: I1125 11:19:59.359520 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" podStartSLOduration=1.6485108560000001 podStartE2EDuration="2.359501751s" podCreationTimestamp="2025-11-25 11:19:57 +0000 UTC" firstStartedPulling="2025-11-25 11:19:58.265305172 +0000 UTC m=+2595.078921759" lastFinishedPulling="2025-11-25 11:19:58.976296067 +0000 UTC m=+2595.789912654" observedRunningTime="2025-11-25 11:19:59.355317278 +0000 UTC m=+2596.168933865" watchObservedRunningTime="2025-11-25 11:19:59.359501751 +0000 UTC m=+2596.173118338" Nov 25 11:20:03 crc kubenswrapper[4696]: I1125 11:20:03.042315 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:20:03 crc kubenswrapper[4696]: E1125 11:20:03.043214 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:20:18 crc kubenswrapper[4696]: I1125 11:20:18.043109 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:20:18 crc kubenswrapper[4696]: E1125 11:20:18.044014 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:20:33 crc kubenswrapper[4696]: I1125 11:20:33.042707 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:20:33 crc kubenswrapper[4696]: E1125 11:20:33.043602 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:20:45 crc kubenswrapper[4696]: I1125 11:20:45.042801 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:20:45 crc kubenswrapper[4696]: E1125 11:20:45.043543 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.530552 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kxgpc"] Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.533734 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.543232 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kxgpc"] Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.641850 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-utilities\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.642060 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-946fs\" (UniqueName: \"kubernetes.io/projected/f94391cb-27f5-494e-ad8e-01e392b5b2c3-kube-api-access-946fs\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.642090 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-catalog-content\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.743739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-946fs\" (UniqueName: \"kubernetes.io/projected/f94391cb-27f5-494e-ad8e-01e392b5b2c3-kube-api-access-946fs\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.743787 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-catalog-content\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.743892 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-utilities\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.744501 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-catalog-content\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.744581 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-utilities\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.765088 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-946fs\" (UniqueName: \"kubernetes.io/projected/f94391cb-27f5-494e-ad8e-01e392b5b2c3-kube-api-access-946fs\") pod \"redhat-marketplace-kxgpc\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:48 crc kubenswrapper[4696]: I1125 11:20:48.858243 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:49 crc kubenswrapper[4696]: I1125 11:20:49.153689 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kxgpc"] Nov 25 11:20:49 crc kubenswrapper[4696]: I1125 11:20:49.780652 4696 generic.go:334] "Generic (PLEG): container finished" podID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerID="043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f" exitCode=0 Nov 25 11:20:49 crc kubenswrapper[4696]: I1125 11:20:49.780857 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerDied","Data":"043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f"} Nov 25 11:20:49 crc kubenswrapper[4696]: I1125 11:20:49.780920 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerStarted","Data":"1ad4913bd6ae973f84c447e814906d5861caefe16609591b71b144f2c5326144"} Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.813108 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerStarted","Data":"49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708"} Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.926920 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lnfvf"] Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.929632 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.947072 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lnfvf"] Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.985026 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-catalog-content\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.985124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j87d7\" (UniqueName: \"kubernetes.io/projected/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-kube-api-access-j87d7\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:50 crc kubenswrapper[4696]: I1125 11:20:50.985333 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-utilities\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.087105 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-utilities\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.087209 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-catalog-content\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.087286 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j87d7\" (UniqueName: \"kubernetes.io/projected/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-kube-api-access-j87d7\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.087651 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-utilities\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.088018 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-catalog-content\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.106866 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j87d7\" (UniqueName: \"kubernetes.io/projected/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-kube-api-access-j87d7\") pod \"community-operators-lnfvf\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.247928 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.803142 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lnfvf"] Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.822490 4696 generic.go:334] "Generic (PLEG): container finished" podID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerID="49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708" exitCode=0 Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.823297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerDied","Data":"49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708"} Nov 25 11:20:51 crc kubenswrapper[4696]: I1125 11:20:51.824897 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerStarted","Data":"9e1e1f3ec2f5faf40f23d687fda527c15d132e7b31a8572eff889d28346b892e"} Nov 25 11:20:52 crc kubenswrapper[4696]: I1125 11:20:52.840325 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerStarted","Data":"34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc"} Nov 25 11:20:52 crc kubenswrapper[4696]: I1125 11:20:52.844583 4696 generic.go:334] "Generic (PLEG): container finished" podID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerID="ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6" exitCode=0 Nov 25 11:20:52 crc kubenswrapper[4696]: I1125 11:20:52.844635 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerDied","Data":"ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6"} Nov 25 11:20:52 crc kubenswrapper[4696]: I1125 11:20:52.868924 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kxgpc" podStartSLOduration=2.41510779 podStartE2EDuration="4.868904375s" podCreationTimestamp="2025-11-25 11:20:48 +0000 UTC" firstStartedPulling="2025-11-25 11:20:49.783428047 +0000 UTC m=+2646.597044644" lastFinishedPulling="2025-11-25 11:20:52.237224632 +0000 UTC m=+2649.050841229" observedRunningTime="2025-11-25 11:20:52.864469895 +0000 UTC m=+2649.678086482" watchObservedRunningTime="2025-11-25 11:20:52.868904375 +0000 UTC m=+2649.682520962" Nov 25 11:20:53 crc kubenswrapper[4696]: I1125 11:20:53.857426 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerStarted","Data":"f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65"} Nov 25 11:20:56 crc kubenswrapper[4696]: I1125 11:20:56.894680 4696 generic.go:334] "Generic (PLEG): container finished" podID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerID="f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65" exitCode=0 Nov 25 11:20:56 crc kubenswrapper[4696]: I1125 11:20:56.894756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerDied","Data":"f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65"} Nov 25 11:20:57 crc kubenswrapper[4696]: I1125 11:20:57.905808 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerStarted","Data":"799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba"} Nov 25 11:20:57 crc kubenswrapper[4696]: I1125 11:20:57.929586 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lnfvf" podStartSLOduration=3.285671123 podStartE2EDuration="7.92956051s" podCreationTimestamp="2025-11-25 11:20:50 +0000 UTC" firstStartedPulling="2025-11-25 11:20:52.847561379 +0000 UTC m=+2649.661177966" lastFinishedPulling="2025-11-25 11:20:57.491450755 +0000 UTC m=+2654.305067353" observedRunningTime="2025-11-25 11:20:57.925922044 +0000 UTC m=+2654.739538651" watchObservedRunningTime="2025-11-25 11:20:57.92956051 +0000 UTC m=+2654.743177097" Nov 25 11:20:58 crc kubenswrapper[4696]: I1125 11:20:58.042201 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:20:58 crc kubenswrapper[4696]: E1125 11:20:58.042463 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:20:58 crc kubenswrapper[4696]: I1125 11:20:58.859354 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:58 crc kubenswrapper[4696]: I1125 11:20:58.859510 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:58 crc kubenswrapper[4696]: I1125 11:20:58.967086 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:20:59 crc kubenswrapper[4696]: I1125 11:20:59.983156 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:21:01 crc kubenswrapper[4696]: I1125 11:21:01.119405 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kxgpc"] Nov 25 11:21:01 crc kubenswrapper[4696]: I1125 11:21:01.248272 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:21:01 crc kubenswrapper[4696]: I1125 11:21:01.248354 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:21:01 crc kubenswrapper[4696]: I1125 11:21:01.297052 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:21:01 crc kubenswrapper[4696]: I1125 11:21:01.942139 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kxgpc" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="registry-server" containerID="cri-o://34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc" gracePeriod=2 Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.386076 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.510117 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-catalog-content\") pod \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.510208 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-utilities\") pod \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.510345 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-946fs\" (UniqueName: \"kubernetes.io/projected/f94391cb-27f5-494e-ad8e-01e392b5b2c3-kube-api-access-946fs\") pod \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\" (UID: \"f94391cb-27f5-494e-ad8e-01e392b5b2c3\") " Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.510979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-utilities" (OuterVolumeSpecName: "utilities") pod "f94391cb-27f5-494e-ad8e-01e392b5b2c3" (UID: "f94391cb-27f5-494e-ad8e-01e392b5b2c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.519022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94391cb-27f5-494e-ad8e-01e392b5b2c3-kube-api-access-946fs" (OuterVolumeSpecName: "kube-api-access-946fs") pod "f94391cb-27f5-494e-ad8e-01e392b5b2c3" (UID: "f94391cb-27f5-494e-ad8e-01e392b5b2c3"). InnerVolumeSpecName "kube-api-access-946fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.524319 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f94391cb-27f5-494e-ad8e-01e392b5b2c3" (UID: "f94391cb-27f5-494e-ad8e-01e392b5b2c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.612602 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-946fs\" (UniqueName: \"kubernetes.io/projected/f94391cb-27f5-494e-ad8e-01e392b5b2c3-kube-api-access-946fs\") on node \"crc\" DevicePath \"\"" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.612637 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.612649 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f94391cb-27f5-494e-ad8e-01e392b5b2c3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.952974 4696 generic.go:334] "Generic (PLEG): container finished" podID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerID="34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc" exitCode=0 Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.953037 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kxgpc" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.953057 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerDied","Data":"34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc"} Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.953434 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kxgpc" event={"ID":"f94391cb-27f5-494e-ad8e-01e392b5b2c3","Type":"ContainerDied","Data":"1ad4913bd6ae973f84c447e814906d5861caefe16609591b71b144f2c5326144"} Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.953482 4696 scope.go:117] "RemoveContainer" containerID="34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.972648 4696 scope.go:117] "RemoveContainer" containerID="49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708" Nov 25 11:21:02 crc kubenswrapper[4696]: I1125 11:21:02.994769 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kxgpc"] Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.004710 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kxgpc"] Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.005381 4696 scope.go:117] "RemoveContainer" containerID="043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f" Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.058474 4696 scope.go:117] "RemoveContainer" containerID="34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc" Nov 25 11:21:03 crc kubenswrapper[4696]: E1125 11:21:03.059131 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc\": container with ID starting with 34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc not found: ID does not exist" containerID="34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc" Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.059177 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc"} err="failed to get container status \"34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc\": rpc error: code = NotFound desc = could not find container \"34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc\": container with ID starting with 34fb436116d1334620b911e399067b0ca383d270800a8e94fe8706ea6b510acc not found: ID does not exist" Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.059206 4696 scope.go:117] "RemoveContainer" containerID="49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708" Nov 25 11:21:03 crc kubenswrapper[4696]: E1125 11:21:03.059729 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708\": container with ID starting with 49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708 not found: ID does not exist" containerID="49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708" Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.059769 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708"} err="failed to get container status \"49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708\": rpc error: code = NotFound desc = could not find container \"49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708\": container with ID starting with 49d7ac0bbda1699609c28df7fe54f9605a8b8af9df59ede401188afefd736708 not found: ID does not exist" Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.059789 4696 scope.go:117] "RemoveContainer" containerID="043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f" Nov 25 11:21:03 crc kubenswrapper[4696]: E1125 11:21:03.060304 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f\": container with ID starting with 043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f not found: ID does not exist" containerID="043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f" Nov 25 11:21:03 crc kubenswrapper[4696]: I1125 11:21:03.060373 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f"} err="failed to get container status \"043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f\": rpc error: code = NotFound desc = could not find container \"043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f\": container with ID starting with 043274f70eca09244bf321c594b0b113dfcba926e169ba9956686bdcf4109c8f not found: ID does not exist" Nov 25 11:21:04 crc kubenswrapper[4696]: I1125 11:21:04.060049 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" path="/var/lib/kubelet/pods/f94391cb-27f5-494e-ad8e-01e392b5b2c3/volumes" Nov 25 11:21:10 crc kubenswrapper[4696]: I1125 11:21:10.042197 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:21:10 crc kubenswrapper[4696]: E1125 11:21:10.042967 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:21:11 crc kubenswrapper[4696]: I1125 11:21:11.298275 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:21:11 crc kubenswrapper[4696]: I1125 11:21:11.350043 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lnfvf"] Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.028828 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lnfvf" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="registry-server" containerID="cri-o://799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba" gracePeriod=2 Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.465054 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.591285 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-catalog-content\") pod \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.591602 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j87d7\" (UniqueName: \"kubernetes.io/projected/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-kube-api-access-j87d7\") pod \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.591862 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-utilities\") pod \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\" (UID: \"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b\") " Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.592499 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-utilities" (OuterVolumeSpecName: "utilities") pod "c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" (UID: "c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.597079 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-kube-api-access-j87d7" (OuterVolumeSpecName: "kube-api-access-j87d7") pod "c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" (UID: "c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b"). InnerVolumeSpecName "kube-api-access-j87d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.640818 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" (UID: "c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.693873 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.693914 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:21:12 crc kubenswrapper[4696]: I1125 11:21:12.693928 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j87d7\" (UniqueName: \"kubernetes.io/projected/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b-kube-api-access-j87d7\") on node \"crc\" DevicePath \"\"" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.039546 4696 generic.go:334] "Generic (PLEG): container finished" podID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerID="799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba" exitCode=0 Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.039595 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnfvf" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.039598 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerDied","Data":"799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba"} Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.039628 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnfvf" event={"ID":"c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b","Type":"ContainerDied","Data":"9e1e1f3ec2f5faf40f23d687fda527c15d132e7b31a8572eff889d28346b892e"} Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.039651 4696 scope.go:117] "RemoveContainer" containerID="799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.084624 4696 scope.go:117] "RemoveContainer" containerID="f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.084958 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lnfvf"] Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.113748 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lnfvf"] Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.142453 4696 scope.go:117] "RemoveContainer" containerID="ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.171041 4696 scope.go:117] "RemoveContainer" containerID="799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba" Nov 25 11:21:13 crc kubenswrapper[4696]: E1125 11:21:13.171562 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba\": container with ID starting with 799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba not found: ID does not exist" containerID="799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.171709 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba"} err="failed to get container status \"799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba\": rpc error: code = NotFound desc = could not find container \"799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba\": container with ID starting with 799d968d8f743fd9ff8de02cdfed37a2e6075f4bf8c408135e7bb165e1ec00ba not found: ID does not exist" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.171797 4696 scope.go:117] "RemoveContainer" containerID="f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65" Nov 25 11:21:13 crc kubenswrapper[4696]: E1125 11:21:13.172960 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65\": container with ID starting with f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65 not found: ID does not exist" containerID="f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.172991 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65"} err="failed to get container status \"f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65\": rpc error: code = NotFound desc = could not find container \"f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65\": container with ID starting with f0f73f2f33a17db41da1ee0e60ecbb843094894849fbdeb977c8e4a04ed84a65 not found: ID does not exist" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.173012 4696 scope.go:117] "RemoveContainer" containerID="ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6" Nov 25 11:21:13 crc kubenswrapper[4696]: E1125 11:21:13.173252 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6\": container with ID starting with ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6 not found: ID does not exist" containerID="ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6" Nov 25 11:21:13 crc kubenswrapper[4696]: I1125 11:21:13.173370 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6"} err="failed to get container status \"ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6\": rpc error: code = NotFound desc = could not find container \"ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6\": container with ID starting with ef9388e6ac2b90c113350dd2af556d7ee7ebb18b51a5ddbdbae7fb7b5cad60c6 not found: ID does not exist" Nov 25 11:21:14 crc kubenswrapper[4696]: I1125 11:21:14.055495 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" path="/var/lib/kubelet/pods/c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b/volumes" Nov 25 11:21:24 crc kubenswrapper[4696]: I1125 11:21:24.049388 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:21:24 crc kubenswrapper[4696]: E1125 11:21:24.050196 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.042015 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.042876 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.075525 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cx59x"] Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.075940 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="registry-server" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.075957 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="registry-server" Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.075973 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="extract-content" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.075981 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="extract-content" Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.076013 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="extract-content" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.076019 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="extract-content" Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.076029 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="registry-server" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.076035 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="registry-server" Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.076046 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="extract-utilities" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.076053 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="extract-utilities" Nov 25 11:21:35 crc kubenswrapper[4696]: E1125 11:21:35.076064 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="extract-utilities" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.076070 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="extract-utilities" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.077124 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94391cb-27f5-494e-ad8e-01e392b5b2c3" containerName="registry-server" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.077190 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3dcdcd5-9dc9-4aec-8a2e-283fc53fcd6b" containerName="registry-server" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.078687 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.101060 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cx59x"] Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.146958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tvl9\" (UniqueName: \"kubernetes.io/projected/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-kube-api-access-5tvl9\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.147216 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-utilities\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.147555 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-catalog-content\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.251493 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-catalog-content\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.251694 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tvl9\" (UniqueName: \"kubernetes.io/projected/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-kube-api-access-5tvl9\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.251730 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-utilities\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.252226 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-catalog-content\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.252595 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-utilities\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.276291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tvl9\" (UniqueName: \"kubernetes.io/projected/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-kube-api-access-5tvl9\") pod \"redhat-operators-cx59x\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.416517 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:35 crc kubenswrapper[4696]: I1125 11:21:35.913438 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cx59x"] Nov 25 11:21:36 crc kubenswrapper[4696]: I1125 11:21:36.237325 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerID="2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522" exitCode=0 Nov 25 11:21:36 crc kubenswrapper[4696]: I1125 11:21:36.237400 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerDied","Data":"2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522"} Nov 25 11:21:36 crc kubenswrapper[4696]: I1125 11:21:36.237634 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerStarted","Data":"87311028aca35314e7cd0839063a18afe121d02f9530b9e05e9faa99139f4c9c"} Nov 25 11:21:38 crc kubenswrapper[4696]: I1125 11:21:38.258070 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerStarted","Data":"deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6"} Nov 25 11:21:44 crc kubenswrapper[4696]: I1125 11:21:44.894197 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:21:44 crc kubenswrapper[4696]: I1125 11:21:44.905878 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:44 crc kubenswrapper[4696]: I1125 11:21:44.909124 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.039445 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-catalog-content\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.039493 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc7xm\" (UniqueName: \"kubernetes.io/projected/7dd837c4-e52e-4b41-857f-6dfa1944693d-kube-api-access-dc7xm\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.039521 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-utilities\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.144365 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-catalog-content\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.144446 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc7xm\" (UniqueName: \"kubernetes.io/projected/7dd837c4-e52e-4b41-857f-6dfa1944693d-kube-api-access-dc7xm\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.144489 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-utilities\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.145894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-utilities\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.146625 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-catalog-content\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.178040 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc7xm\" (UniqueName: \"kubernetes.io/projected/7dd837c4-e52e-4b41-857f-6dfa1944693d-kube-api-access-dc7xm\") pod \"certified-operators-dswtf\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:45 crc kubenswrapper[4696]: I1125 11:21:45.243253 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:21:46 crc kubenswrapper[4696]: I1125 11:21:46.042139 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:21:46 crc kubenswrapper[4696]: E1125 11:21:46.042647 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:21:46 crc kubenswrapper[4696]: I1125 11:21:46.495757 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:21:47 crc kubenswrapper[4696]: I1125 11:21:47.366547 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerID="deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6" exitCode=0 Nov 25 11:21:47 crc kubenswrapper[4696]: I1125 11:21:47.366579 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerDied","Data":"deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6"} Nov 25 11:21:47 crc kubenswrapper[4696]: I1125 11:21:47.368303 4696 generic.go:334] "Generic (PLEG): container finished" podID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerID="fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a" exitCode=0 Nov 25 11:21:47 crc kubenswrapper[4696]: I1125 11:21:47.368557 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dswtf" event={"ID":"7dd837c4-e52e-4b41-857f-6dfa1944693d","Type":"ContainerDied","Data":"fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a"} Nov 25 11:21:47 crc kubenswrapper[4696]: I1125 11:21:47.368580 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dswtf" event={"ID":"7dd837c4-e52e-4b41-857f-6dfa1944693d","Type":"ContainerStarted","Data":"8b9e41d0e8733f5e7fa83d4d17e0d563f01616c4498ac58d52659ec9d5e6dc97"} Nov 25 11:21:49 crc kubenswrapper[4696]: I1125 11:21:49.392095 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerStarted","Data":"f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048"} Nov 25 11:21:49 crc kubenswrapper[4696]: I1125 11:21:49.414970 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cx59x" podStartSLOduration=2.384926933 podStartE2EDuration="14.414952086s" podCreationTimestamp="2025-11-25 11:21:35 +0000 UTC" firstStartedPulling="2025-11-25 11:21:36.2394745 +0000 UTC m=+2693.053091087" lastFinishedPulling="2025-11-25 11:21:48.269499653 +0000 UTC m=+2705.083116240" observedRunningTime="2025-11-25 11:21:49.412469968 +0000 UTC m=+2706.226086555" watchObservedRunningTime="2025-11-25 11:21:49.414952086 +0000 UTC m=+2706.228568673" Nov 25 11:21:54 crc kubenswrapper[4696]: I1125 11:21:54.452989 4696 generic.go:334] "Generic (PLEG): container finished" podID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerID="0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04" exitCode=0 Nov 25 11:21:54 crc kubenswrapper[4696]: I1125 11:21:54.453030 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dswtf" event={"ID":"7dd837c4-e52e-4b41-857f-6dfa1944693d","Type":"ContainerDied","Data":"0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04"} Nov 25 11:21:55 crc kubenswrapper[4696]: I1125 11:21:55.417702 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:55 crc kubenswrapper[4696]: I1125 11:21:55.418972 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:21:55 crc kubenswrapper[4696]: I1125 11:21:55.464747 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dswtf" event={"ID":"7dd837c4-e52e-4b41-857f-6dfa1944693d","Type":"ContainerStarted","Data":"3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a"} Nov 25 11:21:55 crc kubenswrapper[4696]: I1125 11:21:55.492508 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dswtf" podStartSLOduration=3.990840878 podStartE2EDuration="11.492489803s" podCreationTimestamp="2025-11-25 11:21:44 +0000 UTC" firstStartedPulling="2025-11-25 11:21:47.371394683 +0000 UTC m=+2704.185011310" lastFinishedPulling="2025-11-25 11:21:54.873043648 +0000 UTC m=+2711.686660235" observedRunningTime="2025-11-25 11:21:55.485995718 +0000 UTC m=+2712.299612315" watchObservedRunningTime="2025-11-25 11:21:55.492489803 +0000 UTC m=+2712.306106390" Nov 25 11:21:56 crc kubenswrapper[4696]: I1125 11:21:56.503854 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cx59x" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="registry-server" probeResult="failure" output=< Nov 25 11:21:56 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:21:56 crc kubenswrapper[4696]: > Nov 25 11:21:58 crc kubenswrapper[4696]: I1125 11:21:58.042438 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:21:58 crc kubenswrapper[4696]: E1125 11:21:58.043013 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:22:05 crc kubenswrapper[4696]: I1125 11:22:05.244000 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:22:05 crc kubenswrapper[4696]: I1125 11:22:05.244426 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:22:05 crc kubenswrapper[4696]: I1125 11:22:05.308877 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:22:05 crc kubenswrapper[4696]: I1125 11:22:05.602309 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:22:06 crc kubenswrapper[4696]: I1125 11:22:06.133563 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:22:06 crc kubenswrapper[4696]: I1125 11:22:06.275539 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s9cff"] Nov 25 11:22:06 crc kubenswrapper[4696]: I1125 11:22:06.275851 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s9cff" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="registry-server" containerID="cri-o://b47bd19e20eead4570cbc57d0db67874828e25eb19630cea77d764efc212eb59" gracePeriod=2 Nov 25 11:22:06 crc kubenswrapper[4696]: I1125 11:22:06.474199 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cx59x" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="registry-server" probeResult="failure" output=< Nov 25 11:22:06 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:22:06 crc kubenswrapper[4696]: > Nov 25 11:22:06 crc kubenswrapper[4696]: I1125 11:22:06.570176 4696 generic.go:334] "Generic (PLEG): container finished" podID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerID="b47bd19e20eead4570cbc57d0db67874828e25eb19630cea77d764efc212eb59" exitCode=0 Nov 25 11:22:06 crc kubenswrapper[4696]: I1125 11:22:06.570244 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9cff" event={"ID":"de81aa8f-a3f2-42c0-8a6d-1756748ec575","Type":"ContainerDied","Data":"b47bd19e20eead4570cbc57d0db67874828e25eb19630cea77d764efc212eb59"} Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.295079 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.430319 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x696l\" (UniqueName: \"kubernetes.io/projected/de81aa8f-a3f2-42c0-8a6d-1756748ec575-kube-api-access-x696l\") pod \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.430404 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-utilities\") pod \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.430537 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-catalog-content\") pod \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\" (UID: \"de81aa8f-a3f2-42c0-8a6d-1756748ec575\") " Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.433898 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-utilities" (OuterVolumeSpecName: "utilities") pod "de81aa8f-a3f2-42c0-8a6d-1756748ec575" (UID: "de81aa8f-a3f2-42c0-8a6d-1756748ec575"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.448085 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de81aa8f-a3f2-42c0-8a6d-1756748ec575-kube-api-access-x696l" (OuterVolumeSpecName: "kube-api-access-x696l") pod "de81aa8f-a3f2-42c0-8a6d-1756748ec575" (UID: "de81aa8f-a3f2-42c0-8a6d-1756748ec575"). InnerVolumeSpecName "kube-api-access-x696l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.507850 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de81aa8f-a3f2-42c0-8a6d-1756748ec575" (UID: "de81aa8f-a3f2-42c0-8a6d-1756748ec575"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.532925 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x696l\" (UniqueName: \"kubernetes.io/projected/de81aa8f-a3f2-42c0-8a6d-1756748ec575-kube-api-access-x696l\") on node \"crc\" DevicePath \"\"" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.532976 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.532989 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de81aa8f-a3f2-42c0-8a6d-1756748ec575-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.580121 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9cff" event={"ID":"de81aa8f-a3f2-42c0-8a6d-1756748ec575","Type":"ContainerDied","Data":"e9a75391b6383a7a9a21c9771d5dc2e8cdbcd40ccd4cb4d9f3c1c0aac0eccf1e"} Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.580172 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9cff" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.580189 4696 scope.go:117] "RemoveContainer" containerID="b47bd19e20eead4570cbc57d0db67874828e25eb19630cea77d764efc212eb59" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.605615 4696 scope.go:117] "RemoveContainer" containerID="8f5891f5a89280ba498edb75fe8c6a5f8af52c809a58bf341afbb30ab167864e" Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.615137 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s9cff"] Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.622529 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s9cff"] Nov 25 11:22:07 crc kubenswrapper[4696]: I1125 11:22:07.628630 4696 scope.go:117] "RemoveContainer" containerID="7d3d0db5ee1fcae0a8c538889d5767c8ec3f5bfaf96334825adfb498eb2036b6" Nov 25 11:22:08 crc kubenswrapper[4696]: I1125 11:22:08.054724 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" path="/var/lib/kubelet/pods/de81aa8f-a3f2-42c0-8a6d-1756748ec575/volumes" Nov 25 11:22:11 crc kubenswrapper[4696]: I1125 11:22:11.042028 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:22:11 crc kubenswrapper[4696]: I1125 11:22:11.619823 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"6c496966af8a8aab556c6cf92593cbdc2c63bbebab0d4e004fdac5f2f4f8f6cb"} Nov 25 11:22:15 crc kubenswrapper[4696]: I1125 11:22:15.461917 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:22:15 crc kubenswrapper[4696]: I1125 11:22:15.516925 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:22:16 crc kubenswrapper[4696]: I1125 11:22:16.697780 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cx59x"] Nov 25 11:22:16 crc kubenswrapper[4696]: I1125 11:22:16.698314 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cx59x" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="registry-server" containerID="cri-o://f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048" gracePeriod=2 Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.142093 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.226781 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-catalog-content\") pod \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.226908 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-utilities\") pod \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.227713 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-utilities" (OuterVolumeSpecName: "utilities") pod "ff0636c6-c0c3-493d-b2b9-c14a29a79b88" (UID: "ff0636c6-c0c3-493d-b2b9-c14a29a79b88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.227875 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tvl9\" (UniqueName: \"kubernetes.io/projected/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-kube-api-access-5tvl9\") pod \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\" (UID: \"ff0636c6-c0c3-493d-b2b9-c14a29a79b88\") " Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.228706 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.233304 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-kube-api-access-5tvl9" (OuterVolumeSpecName: "kube-api-access-5tvl9") pod "ff0636c6-c0c3-493d-b2b9-c14a29a79b88" (UID: "ff0636c6-c0c3-493d-b2b9-c14a29a79b88"). InnerVolumeSpecName "kube-api-access-5tvl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.320044 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff0636c6-c0c3-493d-b2b9-c14a29a79b88" (UID: "ff0636c6-c0c3-493d-b2b9-c14a29a79b88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.330625 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.330873 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tvl9\" (UniqueName: \"kubernetes.io/projected/ff0636c6-c0c3-493d-b2b9-c14a29a79b88-kube-api-access-5tvl9\") on node \"crc\" DevicePath \"\"" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.673370 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerID="f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048" exitCode=0 Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.673416 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerDied","Data":"f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048"} Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.673450 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cx59x" event={"ID":"ff0636c6-c0c3-493d-b2b9-c14a29a79b88","Type":"ContainerDied","Data":"87311028aca35314e7cd0839063a18afe121d02f9530b9e05e9faa99139f4c9c"} Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.673468 4696 scope.go:117] "RemoveContainer" containerID="f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.673464 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cx59x" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.705043 4696 scope.go:117] "RemoveContainer" containerID="deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.712783 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cx59x"] Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.727836 4696 scope.go:117] "RemoveContainer" containerID="2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.739956 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cx59x"] Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.774791 4696 scope.go:117] "RemoveContainer" containerID="f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048" Nov 25 11:22:17 crc kubenswrapper[4696]: E1125 11:22:17.778229 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048\": container with ID starting with f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048 not found: ID does not exist" containerID="f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.778285 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048"} err="failed to get container status \"f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048\": rpc error: code = NotFound desc = could not find container \"f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048\": container with ID starting with f52ff656e46491cd7b86e1d6e235c4becd0a385b4b20d303ac8c2786ac5de048 not found: ID does not exist" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.778328 4696 scope.go:117] "RemoveContainer" containerID="deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6" Nov 25 11:22:17 crc kubenswrapper[4696]: E1125 11:22:17.778937 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6\": container with ID starting with deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6 not found: ID does not exist" containerID="deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.778960 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6"} err="failed to get container status \"deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6\": rpc error: code = NotFound desc = could not find container \"deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6\": container with ID starting with deeff2f38f31b9b78b9fb4e3997bff4fd8db34cf20189aa05aff377e97e903c6 not found: ID does not exist" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.778974 4696 scope.go:117] "RemoveContainer" containerID="2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522" Nov 25 11:22:17 crc kubenswrapper[4696]: E1125 11:22:17.779351 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522\": container with ID starting with 2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522 not found: ID does not exist" containerID="2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522" Nov 25 11:22:17 crc kubenswrapper[4696]: I1125 11:22:17.779378 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522"} err="failed to get container status \"2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522\": rpc error: code = NotFound desc = could not find container \"2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522\": container with ID starting with 2b9a7b139d5806787ebf6e922d53f5fb0dad411eae0854436cc184ffad0b3522 not found: ID does not exist" Nov 25 11:22:18 crc kubenswrapper[4696]: I1125 11:22:18.052396 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" path="/var/lib/kubelet/pods/ff0636c6-c0c3-493d-b2b9-c14a29a79b88/volumes" Nov 25 11:23:15 crc kubenswrapper[4696]: I1125 11:23:15.192099 4696 generic.go:334] "Generic (PLEG): container finished" podID="11902ad3-7187-4749-8fbf-3d5f1eb77764" containerID="68259a8a34701b40f657503776589fd7374cc2fbf65415eaac39e1d514760dd2" exitCode=0 Nov 25 11:23:15 crc kubenswrapper[4696]: I1125 11:23:15.192184 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" event={"ID":"11902ad3-7187-4749-8fbf-3d5f1eb77764","Type":"ContainerDied","Data":"68259a8a34701b40f657503776589fd7374cc2fbf65415eaac39e1d514760dd2"} Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.595629 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752095 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-combined-ca-bundle\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752153 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-ssh-key\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752202 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-inventory\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752314 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-1\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752341 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf6kp\" (UniqueName: \"kubernetes.io/projected/11902ad3-7187-4749-8fbf-3d5f1eb77764-kube-api-access-rf6kp\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752364 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-extra-config-0\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752428 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-1\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752475 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-0\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.752531 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-0\") pod \"11902ad3-7187-4749-8fbf-3d5f1eb77764\" (UID: \"11902ad3-7187-4749-8fbf-3d5f1eb77764\") " Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.759425 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.760228 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11902ad3-7187-4749-8fbf-3d5f1eb77764-kube-api-access-rf6kp" (OuterVolumeSpecName: "kube-api-access-rf6kp") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "kube-api-access-rf6kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.782570 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.783874 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.784095 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.794323 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-inventory" (OuterVolumeSpecName: "inventory") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.794436 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.806856 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.808345 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "11902ad3-7187-4749-8fbf-3d5f1eb77764" (UID: "11902ad3-7187-4749-8fbf-3d5f1eb77764"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855288 4696 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855323 4696 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855352 4696 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855362 4696 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855371 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855379 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855388 4696 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855430 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf6kp\" (UniqueName: \"kubernetes.io/projected/11902ad3-7187-4749-8fbf-3d5f1eb77764-kube-api-access-rf6kp\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:16 crc kubenswrapper[4696]: I1125 11:23:16.855439 4696 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/11902ad3-7187-4749-8fbf-3d5f1eb77764-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.214570 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" event={"ID":"11902ad3-7187-4749-8fbf-3d5f1eb77764","Type":"ContainerDied","Data":"fd427399fc2cdb7de2b1bda64803fc84dfd63c2ebb720c0065f08d246c5abb65"} Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.214612 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd427399fc2cdb7de2b1bda64803fc84dfd63c2ebb720c0065f08d246c5abb65" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.214621 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fkn4" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379342 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz"] Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379770 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="registry-server" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379803 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="registry-server" Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379812 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11902ad3-7187-4749-8fbf-3d5f1eb77764" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379819 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="11902ad3-7187-4749-8fbf-3d5f1eb77764" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379831 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="registry-server" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379838 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="registry-server" Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379850 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="extract-utilities" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379856 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="extract-utilities" Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379875 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="extract-utilities" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379881 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="extract-utilities" Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379896 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="extract-content" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379902 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="extract-content" Nov 25 11:23:17 crc kubenswrapper[4696]: E1125 11:23:17.379912 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="extract-content" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.379917 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="extract-content" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.380096 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="11902ad3-7187-4749-8fbf-3d5f1eb77764" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.380124 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="de81aa8f-a3f2-42c0-8a6d-1756748ec575" containerName="registry-server" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.380138 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0636c6-c0c3-493d-b2b9-c14a29a79b88" containerName="registry-server" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.380812 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.383136 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.383140 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-77nt5" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.383140 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.387800 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.387800 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.401660 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz"] Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.467259 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.467624 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwpsj\" (UniqueName: \"kubernetes.io/projected/372198a7-8311-4133-8047-fcb9908b7f90-kube-api-access-nwpsj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.467835 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.467960 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.468177 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.468364 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.468553 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570209 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570317 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570368 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570391 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570413 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwpsj\" (UniqueName: \"kubernetes.io/projected/372198a7-8311-4133-8047-fcb9908b7f90-kube-api-access-nwpsj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570456 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.570477 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.574919 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.575180 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.576152 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.576393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.576616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.577003 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.589964 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwpsj\" (UniqueName: \"kubernetes.io/projected/372198a7-8311-4133-8047-fcb9908b7f90-kube-api-access-nwpsj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:17 crc kubenswrapper[4696]: I1125 11:23:17.701044 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:23:18 crc kubenswrapper[4696]: I1125 11:23:18.218176 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz"] Nov 25 11:23:19 crc kubenswrapper[4696]: I1125 11:23:19.235887 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" event={"ID":"372198a7-8311-4133-8047-fcb9908b7f90","Type":"ContainerStarted","Data":"85432f3bb82bb8c347424845613e41c8924acb835126669d47392c849ae545ab"} Nov 25 11:23:19 crc kubenswrapper[4696]: I1125 11:23:19.237719 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" event={"ID":"372198a7-8311-4133-8047-fcb9908b7f90","Type":"ContainerStarted","Data":"7c2a955029c906a7a5532015428d8596752d7e1c946e74f3786e4ff9ad51afe1"} Nov 25 11:23:19 crc kubenswrapper[4696]: I1125 11:23:19.258101 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" podStartSLOduration=1.604150691 podStartE2EDuration="2.25808313s" podCreationTimestamp="2025-11-25 11:23:17 +0000 UTC" firstStartedPulling="2025-11-25 11:23:18.22090727 +0000 UTC m=+2795.034523857" lastFinishedPulling="2025-11-25 11:23:18.874839709 +0000 UTC m=+2795.688456296" observedRunningTime="2025-11-25 11:23:19.254512458 +0000 UTC m=+2796.068129045" watchObservedRunningTime="2025-11-25 11:23:19.25808313 +0000 UTC m=+2796.071699717" Nov 25 11:24:30 crc kubenswrapper[4696]: I1125 11:24:30.802189 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:24:30 crc kubenswrapper[4696]: I1125 11:24:30.802832 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:25:00 crc kubenswrapper[4696]: I1125 11:25:00.801645 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:25:00 crc kubenswrapper[4696]: I1125 11:25:00.802839 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:25:30 crc kubenswrapper[4696]: I1125 11:25:30.801909 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:25:30 crc kubenswrapper[4696]: I1125 11:25:30.802445 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:25:30 crc kubenswrapper[4696]: I1125 11:25:30.802492 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:25:30 crc kubenswrapper[4696]: I1125 11:25:30.803276 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c496966af8a8aab556c6cf92593cbdc2c63bbebab0d4e004fdac5f2f4f8f6cb"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:25:30 crc kubenswrapper[4696]: I1125 11:25:30.803331 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://6c496966af8a8aab556c6cf92593cbdc2c63bbebab0d4e004fdac5f2f4f8f6cb" gracePeriod=600 Nov 25 11:25:31 crc kubenswrapper[4696]: I1125 11:25:31.449107 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="6c496966af8a8aab556c6cf92593cbdc2c63bbebab0d4e004fdac5f2f4f8f6cb" exitCode=0 Nov 25 11:25:31 crc kubenswrapper[4696]: I1125 11:25:31.449385 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"6c496966af8a8aab556c6cf92593cbdc2c63bbebab0d4e004fdac5f2f4f8f6cb"} Nov 25 11:25:31 crc kubenswrapper[4696]: I1125 11:25:31.449934 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83"} Nov 25 11:25:31 crc kubenswrapper[4696]: I1125 11:25:31.449959 4696 scope.go:117] "RemoveContainer" containerID="280cdb74b05b4862017dbd0395adc79436b48368908df140e7102f1b20621c09" Nov 25 11:26:44 crc kubenswrapper[4696]: I1125 11:26:44.072935 4696 generic.go:334] "Generic (PLEG): container finished" podID="372198a7-8311-4133-8047-fcb9908b7f90" containerID="85432f3bb82bb8c347424845613e41c8924acb835126669d47392c849ae545ab" exitCode=0 Nov 25 11:26:44 crc kubenswrapper[4696]: I1125 11:26:44.073076 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" event={"ID":"372198a7-8311-4133-8047-fcb9908b7f90","Type":"ContainerDied","Data":"85432f3bb82bb8c347424845613e41c8924acb835126669d47392c849ae545ab"} Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.517616 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697231 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-inventory\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697409 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ssh-key\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697463 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwpsj\" (UniqueName: \"kubernetes.io/projected/372198a7-8311-4133-8047-fcb9908b7f90-kube-api-access-nwpsj\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697522 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-1\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697561 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-0\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697798 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-2\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.697861 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-telemetry-combined-ca-bundle\") pod \"372198a7-8311-4133-8047-fcb9908b7f90\" (UID: \"372198a7-8311-4133-8047-fcb9908b7f90\") " Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.703450 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.705967 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/372198a7-8311-4133-8047-fcb9908b7f90-kube-api-access-nwpsj" (OuterVolumeSpecName: "kube-api-access-nwpsj") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "kube-api-access-nwpsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.727470 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.729580 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.731601 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.732126 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-inventory" (OuterVolumeSpecName: "inventory") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.739822 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "372198a7-8311-4133-8047-fcb9908b7f90" (UID: "372198a7-8311-4133-8047-fcb9908b7f90"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803678 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803721 4696 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803742 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803753 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803766 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwpsj\" (UniqueName: \"kubernetes.io/projected/372198a7-8311-4133-8047-fcb9908b7f90-kube-api-access-nwpsj\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803776 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:45 crc kubenswrapper[4696]: I1125 11:26:45.803791 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/372198a7-8311-4133-8047-fcb9908b7f90-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:46 crc kubenswrapper[4696]: I1125 11:26:46.090955 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" event={"ID":"372198a7-8311-4133-8047-fcb9908b7f90","Type":"ContainerDied","Data":"7c2a955029c906a7a5532015428d8596752d7e1c946e74f3786e4ff9ad51afe1"} Nov 25 11:26:46 crc kubenswrapper[4696]: I1125 11:26:46.090998 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c2a955029c906a7a5532015428d8596752d7e1c946e74f3786e4ff9ad51afe1" Nov 25 11:26:46 crc kubenswrapper[4696]: I1125 11:26:46.091080 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.702327 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 11:27:44 crc kubenswrapper[4696]: E1125 11:27:44.703355 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="372198a7-8311-4133-8047-fcb9908b7f90" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.703376 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="372198a7-8311-4133-8047-fcb9908b7f90" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.703628 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="372198a7-8311-4133-8047-fcb9908b7f90" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.704597 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.717555 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.717656 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.717714 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.720046 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-58qxz" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.750107 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805586 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805623 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805646 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805705 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805764 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805785 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805812 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805827 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5tzh\" (UniqueName: \"kubernetes.io/projected/f8d03f21-fad7-432e-8220-12094b55a1a1-kube-api-access-d5tzh\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.805866 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-config-data\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907268 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907384 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907417 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907457 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907480 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5tzh\" (UniqueName: \"kubernetes.io/projected/f8d03f21-fad7-432e-8220-12094b55a1a1-kube-api-access-d5tzh\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907536 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-config-data\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907612 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907637 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.907830 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.909135 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-config-data\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.909894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.911164 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.911872 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.914386 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.936598 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.943409 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.944082 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5tzh\" (UniqueName: \"kubernetes.io/projected/f8d03f21-fad7-432e-8220-12094b55a1a1-kube-api-access-d5tzh\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:44 crc kubenswrapper[4696]: I1125 11:27:44.997925 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " pod="openstack/tempest-tests-tempest" Nov 25 11:27:45 crc kubenswrapper[4696]: I1125 11:27:45.037827 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:27:45 crc kubenswrapper[4696]: I1125 11:27:45.494564 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:27:45 crc kubenswrapper[4696]: I1125 11:27:45.498324 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 11:27:45 crc kubenswrapper[4696]: I1125 11:27:45.630458 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f8d03f21-fad7-432e-8220-12094b55a1a1","Type":"ContainerStarted","Data":"3e59e4a28d02156283086c974188223877485ed0006fe7d6e50cdf162a345764"} Nov 25 11:28:00 crc kubenswrapper[4696]: I1125 11:28:00.802340 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:28:00 crc kubenswrapper[4696]: I1125 11:28:00.802930 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:28:20 crc kubenswrapper[4696]: E1125 11:28:20.898091 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 25 11:28:20 crc kubenswrapper[4696]: E1125 11:28:20.898762 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d5tzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(f8d03f21-fad7-432e-8220-12094b55a1a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 11:28:20 crc kubenswrapper[4696]: E1125 11:28:20.899989 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="f8d03f21-fad7-432e-8220-12094b55a1a1" Nov 25 11:28:21 crc kubenswrapper[4696]: E1125 11:28:21.015764 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="f8d03f21-fad7-432e-8220-12094b55a1a1" Nov 25 11:28:30 crc kubenswrapper[4696]: I1125 11:28:30.802582 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:28:30 crc kubenswrapper[4696]: I1125 11:28:30.803119 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:28:36 crc kubenswrapper[4696]: I1125 11:28:36.148371 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f8d03f21-fad7-432e-8220-12094b55a1a1","Type":"ContainerStarted","Data":"7a2a28c3619519f7d9e3fc6ad12511859129535bb6d30665e6fd9690f9995610"} Nov 25 11:28:36 crc kubenswrapper[4696]: I1125 11:28:36.172896 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.847505367 podStartE2EDuration="53.172874537s" podCreationTimestamp="2025-11-25 11:27:43 +0000 UTC" firstStartedPulling="2025-11-25 11:27:45.494369918 +0000 UTC m=+3062.307986505" lastFinishedPulling="2025-11-25 11:28:34.819739088 +0000 UTC m=+3111.633355675" observedRunningTime="2025-11-25 11:28:36.167277 +0000 UTC m=+3112.980893597" watchObservedRunningTime="2025-11-25 11:28:36.172874537 +0000 UTC m=+3112.986491124" Nov 25 11:29:00 crc kubenswrapper[4696]: I1125 11:29:00.802474 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:29:00 crc kubenswrapper[4696]: I1125 11:29:00.802904 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:29:00 crc kubenswrapper[4696]: I1125 11:29:00.802946 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:29:00 crc kubenswrapper[4696]: I1125 11:29:00.803587 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:29:00 crc kubenswrapper[4696]: I1125 11:29:00.803635 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" gracePeriod=600 Nov 25 11:29:00 crc kubenswrapper[4696]: E1125 11:29:00.931763 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:29:01 crc kubenswrapper[4696]: I1125 11:29:01.373901 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" exitCode=0 Nov 25 11:29:01 crc kubenswrapper[4696]: I1125 11:29:01.373946 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83"} Nov 25 11:29:01 crc kubenswrapper[4696]: I1125 11:29:01.373996 4696 scope.go:117] "RemoveContainer" containerID="6c496966af8a8aab556c6cf92593cbdc2c63bbebab0d4e004fdac5f2f4f8f6cb" Nov 25 11:29:01 crc kubenswrapper[4696]: I1125 11:29:01.374706 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:29:01 crc kubenswrapper[4696]: E1125 11:29:01.374982 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:29:14 crc kubenswrapper[4696]: I1125 11:29:14.049276 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:29:14 crc kubenswrapper[4696]: E1125 11:29:14.050125 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:29:29 crc kubenswrapper[4696]: I1125 11:29:29.042320 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:29:29 crc kubenswrapper[4696]: E1125 11:29:29.042957 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:29:44 crc kubenswrapper[4696]: I1125 11:29:44.053184 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:29:44 crc kubenswrapper[4696]: E1125 11:29:44.054091 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:29:55 crc kubenswrapper[4696]: I1125 11:29:55.042229 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:29:55 crc kubenswrapper[4696]: E1125 11:29:55.043078 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.180064 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj"] Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.182123 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.184367 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.205916 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.251555 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj"] Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.327999 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-config-volume\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.328062 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-secret-volume\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.328108 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67zfh\" (UniqueName: \"kubernetes.io/projected/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-kube-api-access-67zfh\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.429880 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-config-volume\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.429987 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-secret-volume\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.430093 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67zfh\" (UniqueName: \"kubernetes.io/projected/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-kube-api-access-67zfh\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.430751 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-config-volume\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.449769 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-secret-volume\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.450562 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67zfh\" (UniqueName: \"kubernetes.io/projected/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-kube-api-access-67zfh\") pod \"collect-profiles-29401170-w6ncj\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:00 crc kubenswrapper[4696]: I1125 11:30:00.520204 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:01 crc kubenswrapper[4696]: I1125 11:30:01.366150 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj"] Nov 25 11:30:01 crc kubenswrapper[4696]: I1125 11:30:01.898639 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" event={"ID":"2f4cff06-b620-4a8f-b4d3-fc02eae849ed","Type":"ContainerStarted","Data":"c881a15d1734faf0b3f9031b1747f6b6371fc38ca0f37dcef0f3a573a8d2355e"} Nov 25 11:30:01 crc kubenswrapper[4696]: I1125 11:30:01.899038 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" event={"ID":"2f4cff06-b620-4a8f-b4d3-fc02eae849ed","Type":"ContainerStarted","Data":"6fd3183c41490abbd89b91e507c45b6d68f6eb810fcdec5a7a71bda4cbbf09cd"} Nov 25 11:30:02 crc kubenswrapper[4696]: I1125 11:30:02.908954 4696 generic.go:334] "Generic (PLEG): container finished" podID="2f4cff06-b620-4a8f-b4d3-fc02eae849ed" containerID="c881a15d1734faf0b3f9031b1747f6b6371fc38ca0f37dcef0f3a573a8d2355e" exitCode=0 Nov 25 11:30:02 crc kubenswrapper[4696]: I1125 11:30:02.909267 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" event={"ID":"2f4cff06-b620-4a8f-b4d3-fc02eae849ed","Type":"ContainerDied","Data":"c881a15d1734faf0b3f9031b1747f6b6371fc38ca0f37dcef0f3a573a8d2355e"} Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.418111 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.508104 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67zfh\" (UniqueName: \"kubernetes.io/projected/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-kube-api-access-67zfh\") pod \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.508348 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-secret-volume\") pod \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.508427 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-config-volume\") pod \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\" (UID: \"2f4cff06-b620-4a8f-b4d3-fc02eae849ed\") " Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.509220 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f4cff06-b620-4a8f-b4d3-fc02eae849ed" (UID: "2f4cff06-b620-4a8f-b4d3-fc02eae849ed"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.521515 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f4cff06-b620-4a8f-b4d3-fc02eae849ed" (UID: "2f4cff06-b620-4a8f-b4d3-fc02eae849ed"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.521639 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-kube-api-access-67zfh" (OuterVolumeSpecName: "kube-api-access-67zfh") pod "2f4cff06-b620-4a8f-b4d3-fc02eae849ed" (UID: "2f4cff06-b620-4a8f-b4d3-fc02eae849ed"). InnerVolumeSpecName "kube-api-access-67zfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.611113 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.611150 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67zfh\" (UniqueName: \"kubernetes.io/projected/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-kube-api-access-67zfh\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.611160 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f4cff06-b620-4a8f-b4d3-fc02eae849ed-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.925555 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" event={"ID":"2f4cff06-b620-4a8f-b4d3-fc02eae849ed","Type":"ContainerDied","Data":"6fd3183c41490abbd89b91e507c45b6d68f6eb810fcdec5a7a71bda4cbbf09cd"} Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.925602 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fd3183c41490abbd89b91e507c45b6d68f6eb810fcdec5a7a71bda4cbbf09cd" Nov 25 11:30:04 crc kubenswrapper[4696]: I1125 11:30:04.925867 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-w6ncj" Nov 25 11:30:05 crc kubenswrapper[4696]: I1125 11:30:05.504622 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4"] Nov 25 11:30:05 crc kubenswrapper[4696]: I1125 11:30:05.514554 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hldk4"] Nov 25 11:30:06 crc kubenswrapper[4696]: I1125 11:30:06.057920 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f046cd-d8e3-4ee3-aa07-2ea312ccf224" path="/var/lib/kubelet/pods/59f046cd-d8e3-4ee3-aa07-2ea312ccf224/volumes" Nov 25 11:30:07 crc kubenswrapper[4696]: I1125 11:30:07.042691 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:30:07 crc kubenswrapper[4696]: E1125 11:30:07.043011 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:30:18 crc kubenswrapper[4696]: I1125 11:30:18.041889 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:30:18 crc kubenswrapper[4696]: E1125 11:30:18.042620 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:30:30 crc kubenswrapper[4696]: I1125 11:30:30.043213 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:30:30 crc kubenswrapper[4696]: E1125 11:30:30.043858 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:30:36 crc kubenswrapper[4696]: I1125 11:30:36.865258 4696 scope.go:117] "RemoveContainer" containerID="0c35d6b79f6ab2ca262279badacf8b0bae6e8d68bc002fe88eb656634e5cecb8" Nov 25 11:30:42 crc kubenswrapper[4696]: I1125 11:30:42.042444 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:30:42 crc kubenswrapper[4696]: E1125 11:30:42.043202 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:30:53 crc kubenswrapper[4696]: I1125 11:30:53.061899 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:30:53 crc kubenswrapper[4696]: E1125 11:30:53.062616 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:31:08 crc kubenswrapper[4696]: I1125 11:31:08.043202 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:31:08 crc kubenswrapper[4696]: E1125 11:31:08.044019 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:31:20 crc kubenswrapper[4696]: I1125 11:31:20.042550 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:31:20 crc kubenswrapper[4696]: E1125 11:31:20.043462 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:31:32 crc kubenswrapper[4696]: I1125 11:31:32.042079 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:31:32 crc kubenswrapper[4696]: E1125 11:31:32.042860 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:31:47 crc kubenswrapper[4696]: I1125 11:31:47.042551 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:31:47 crc kubenswrapper[4696]: E1125 11:31:47.043451 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.672989 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xnhp6"] Nov 25 11:31:48 crc kubenswrapper[4696]: E1125 11:31:48.673779 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4cff06-b620-4a8f-b4d3-fc02eae849ed" containerName="collect-profiles" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.673798 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4cff06-b620-4a8f-b4d3-fc02eae849ed" containerName="collect-profiles" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.674048 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4cff06-b620-4a8f-b4d3-fc02eae849ed" containerName="collect-profiles" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.675617 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.699620 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnhp6"] Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.792276 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-catalog-content\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.792342 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x6hd\" (UniqueName: \"kubernetes.io/projected/32f150fe-615d-4aa9-ae0c-47059d5714a0-kube-api-access-8x6hd\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.792573 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-utilities\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.894847 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-catalog-content\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.894919 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x6hd\" (UniqueName: \"kubernetes.io/projected/32f150fe-615d-4aa9-ae0c-47059d5714a0-kube-api-access-8x6hd\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.894970 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-utilities\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.895366 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-catalog-content\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.895445 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-utilities\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.918226 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x6hd\" (UniqueName: \"kubernetes.io/projected/32f150fe-615d-4aa9-ae0c-47059d5714a0-kube-api-access-8x6hd\") pod \"redhat-marketplace-xnhp6\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:48 crc kubenswrapper[4696]: I1125 11:31:48.993723 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:49 crc kubenswrapper[4696]: I1125 11:31:49.608017 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnhp6"] Nov 25 11:31:49 crc kubenswrapper[4696]: I1125 11:31:49.855129 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerStarted","Data":"9b947c0be7952509efbbfb01ea32a5ca71a322b35a5504b307ffc5c11a59dca7"} Nov 25 11:31:50 crc kubenswrapper[4696]: I1125 11:31:50.867680 4696 generic.go:334] "Generic (PLEG): container finished" podID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerID="fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e" exitCode=0 Nov 25 11:31:50 crc kubenswrapper[4696]: I1125 11:31:50.867741 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerDied","Data":"fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e"} Nov 25 11:31:52 crc kubenswrapper[4696]: I1125 11:31:52.898756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerStarted","Data":"fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141"} Nov 25 11:31:53 crc kubenswrapper[4696]: I1125 11:31:53.909041 4696 generic.go:334] "Generic (PLEG): container finished" podID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerID="fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141" exitCode=0 Nov 25 11:31:53 crc kubenswrapper[4696]: I1125 11:31:53.909378 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerDied","Data":"fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141"} Nov 25 11:31:54 crc kubenswrapper[4696]: I1125 11:31:54.919950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerStarted","Data":"34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c"} Nov 25 11:31:54 crc kubenswrapper[4696]: I1125 11:31:54.944588 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xnhp6" podStartSLOduration=3.428148455 podStartE2EDuration="6.944562896s" podCreationTimestamp="2025-11-25 11:31:48 +0000 UTC" firstStartedPulling="2025-11-25 11:31:50.871169233 +0000 UTC m=+3307.684785820" lastFinishedPulling="2025-11-25 11:31:54.387583674 +0000 UTC m=+3311.201200261" observedRunningTime="2025-11-25 11:31:54.936368805 +0000 UTC m=+3311.749985412" watchObservedRunningTime="2025-11-25 11:31:54.944562896 +0000 UTC m=+3311.758179483" Nov 25 11:31:58 crc kubenswrapper[4696]: I1125 11:31:58.997088 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:58 crc kubenswrapper[4696]: I1125 11:31:58.997469 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:31:59 crc kubenswrapper[4696]: I1125 11:31:59.042888 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:31:59 crc kubenswrapper[4696]: E1125 11:31:59.043697 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:31:59 crc kubenswrapper[4696]: I1125 11:31:59.049276 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:32:00 crc kubenswrapper[4696]: I1125 11:32:00.018111 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:32:00 crc kubenswrapper[4696]: I1125 11:32:00.077132 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnhp6"] Nov 25 11:32:01 crc kubenswrapper[4696]: I1125 11:32:01.983441 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xnhp6" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="registry-server" containerID="cri-o://34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c" gracePeriod=2 Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.595703 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.773401 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-utilities\") pod \"32f150fe-615d-4aa9-ae0c-47059d5714a0\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.773481 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x6hd\" (UniqueName: \"kubernetes.io/projected/32f150fe-615d-4aa9-ae0c-47059d5714a0-kube-api-access-8x6hd\") pod \"32f150fe-615d-4aa9-ae0c-47059d5714a0\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.773523 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-catalog-content\") pod \"32f150fe-615d-4aa9-ae0c-47059d5714a0\" (UID: \"32f150fe-615d-4aa9-ae0c-47059d5714a0\") " Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.775022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-utilities" (OuterVolumeSpecName: "utilities") pod "32f150fe-615d-4aa9-ae0c-47059d5714a0" (UID: "32f150fe-615d-4aa9-ae0c-47059d5714a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.791622 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32f150fe-615d-4aa9-ae0c-47059d5714a0-kube-api-access-8x6hd" (OuterVolumeSpecName: "kube-api-access-8x6hd") pod "32f150fe-615d-4aa9-ae0c-47059d5714a0" (UID: "32f150fe-615d-4aa9-ae0c-47059d5714a0"). InnerVolumeSpecName "kube-api-access-8x6hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.796908 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32f150fe-615d-4aa9-ae0c-47059d5714a0" (UID: "32f150fe-615d-4aa9-ae0c-47059d5714a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.876094 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.876134 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x6hd\" (UniqueName: \"kubernetes.io/projected/32f150fe-615d-4aa9-ae0c-47059d5714a0-kube-api-access-8x6hd\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.876147 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32f150fe-615d-4aa9-ae0c-47059d5714a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.994304 4696 generic.go:334] "Generic (PLEG): container finished" podID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerID="34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c" exitCode=0 Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.994359 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnhp6" Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.994439 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerDied","Data":"34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c"} Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.995626 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnhp6" event={"ID":"32f150fe-615d-4aa9-ae0c-47059d5714a0","Type":"ContainerDied","Data":"9b947c0be7952509efbbfb01ea32a5ca71a322b35a5504b307ffc5c11a59dca7"} Nov 25 11:32:02 crc kubenswrapper[4696]: I1125 11:32:02.995653 4696 scope.go:117] "RemoveContainer" containerID="34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.020689 4696 scope.go:117] "RemoveContainer" containerID="fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.033758 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnhp6"] Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.045925 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnhp6"] Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.060876 4696 scope.go:117] "RemoveContainer" containerID="fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.111295 4696 scope.go:117] "RemoveContainer" containerID="34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c" Nov 25 11:32:03 crc kubenswrapper[4696]: E1125 11:32:03.111890 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c\": container with ID starting with 34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c not found: ID does not exist" containerID="34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.111936 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c"} err="failed to get container status \"34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c\": rpc error: code = NotFound desc = could not find container \"34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c\": container with ID starting with 34424b35cb66f3e77aa4e9bb79f228397cef51abd173245d22c69d1792b6691c not found: ID does not exist" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.111963 4696 scope.go:117] "RemoveContainer" containerID="fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141" Nov 25 11:32:03 crc kubenswrapper[4696]: E1125 11:32:03.112387 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141\": container with ID starting with fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141 not found: ID does not exist" containerID="fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.112426 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141"} err="failed to get container status \"fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141\": rpc error: code = NotFound desc = could not find container \"fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141\": container with ID starting with fcd66c6d9a1e9480cc0f1f6d8d9f138a6090667f06c3a4d401d83b0fec87e141 not found: ID does not exist" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.112451 4696 scope.go:117] "RemoveContainer" containerID="fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e" Nov 25 11:32:03 crc kubenswrapper[4696]: E1125 11:32:03.112884 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e\": container with ID starting with fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e not found: ID does not exist" containerID="fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.112923 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e"} err="failed to get container status \"fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e\": rpc error: code = NotFound desc = could not find container \"fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e\": container with ID starting with fbb10b38ce35bbf53bcffa58cf2e73e8264b47b87d7363eea5d8894df88f2b8e not found: ID does not exist" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.707160 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:32:03 crc kubenswrapper[4696]: E1125 11:32:03.707629 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="extract-utilities" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.707655 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="extract-utilities" Nov 25 11:32:03 crc kubenswrapper[4696]: E1125 11:32:03.707710 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="registry-server" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.707718 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="registry-server" Nov 25 11:32:03 crc kubenswrapper[4696]: E1125 11:32:03.707732 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="extract-content" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.707740 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="extract-content" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.707982 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" containerName="registry-server" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.710710 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.756054 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.807811 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-utilities\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.808120 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkm2w\" (UniqueName: \"kubernetes.io/projected/a4329258-c066-423c-b084-db2267dddec8-kube-api-access-mkm2w\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.808232 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-catalog-content\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.909561 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-catalog-content\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.909654 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-utilities\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.909763 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkm2w\" (UniqueName: \"kubernetes.io/projected/a4329258-c066-423c-b084-db2267dddec8-kube-api-access-mkm2w\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.910251 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-catalog-content\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.910284 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-utilities\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:03 crc kubenswrapper[4696]: I1125 11:32:03.935450 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkm2w\" (UniqueName: \"kubernetes.io/projected/a4329258-c066-423c-b084-db2267dddec8-kube-api-access-mkm2w\") pod \"redhat-operators-2jlsf\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:04 crc kubenswrapper[4696]: I1125 11:32:04.043893 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:04 crc kubenswrapper[4696]: I1125 11:32:04.053924 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32f150fe-615d-4aa9-ae0c-47059d5714a0" path="/var/lib/kubelet/pods/32f150fe-615d-4aa9-ae0c-47059d5714a0/volumes" Nov 25 11:32:04 crc kubenswrapper[4696]: I1125 11:32:04.598378 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.014497 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4329258-c066-423c-b084-db2267dddec8" containerID="8b4926f1a525b52937e1c90ddead1b04a851f6ae6ab96a641245873ce715baaa" exitCode=0 Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.014537 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerDied","Data":"8b4926f1a525b52937e1c90ddead1b04a851f6ae6ab96a641245873ce715baaa"} Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.014817 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerStarted","Data":"e3bcb41e1eeb20dbd59e754eed34a15d72207410d4b69f8dea04a884229534ea"} Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.307649 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9hlrx"] Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.312231 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.338617 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlr9f\" (UniqueName: \"kubernetes.io/projected/25e551df-fa3d-40bd-b2e0-e972976f5c74-kube-api-access-vlr9f\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.338650 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e551df-fa3d-40bd-b2e0-e972976f5c74-utilities\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.338691 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e551df-fa3d-40bd-b2e0-e972976f5c74-catalog-content\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.346489 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9hlrx"] Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.440201 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e551df-fa3d-40bd-b2e0-e972976f5c74-catalog-content\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.440493 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlr9f\" (UniqueName: \"kubernetes.io/projected/25e551df-fa3d-40bd-b2e0-e972976f5c74-kube-api-access-vlr9f\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.440525 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e551df-fa3d-40bd-b2e0-e972976f5c74-utilities\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.441121 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e551df-fa3d-40bd-b2e0-e972976f5c74-utilities\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.441421 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e551df-fa3d-40bd-b2e0-e972976f5c74-catalog-content\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.487579 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlr9f\" (UniqueName: \"kubernetes.io/projected/25e551df-fa3d-40bd-b2e0-e972976f5c74-kube-api-access-vlr9f\") pod \"certified-operators-9hlrx\" (UID: \"25e551df-fa3d-40bd-b2e0-e972976f5c74\") " pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:05 crc kubenswrapper[4696]: I1125 11:32:05.664523 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:06 crc kubenswrapper[4696]: I1125 11:32:06.421094 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9hlrx"] Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.070351 4696 generic.go:334] "Generic (PLEG): container finished" podID="25e551df-fa3d-40bd-b2e0-e972976f5c74" containerID="a0ecdd11d8619915056e2e68ae92171f0039171a25604ef0f2411eb6070fae76" exitCode=0 Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.070701 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hlrx" event={"ID":"25e551df-fa3d-40bd-b2e0-e972976f5c74","Type":"ContainerDied","Data":"a0ecdd11d8619915056e2e68ae92171f0039171a25604ef0f2411eb6070fae76"} Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.070944 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hlrx" event={"ID":"25e551df-fa3d-40bd-b2e0-e972976f5c74","Type":"ContainerStarted","Data":"db0a3a49bfd1c23124084a4b08e960dcdfe4ae2fb3d9609fd73e5efda218d5f1"} Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.135619 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-64lbd"] Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.138279 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.171212 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-64lbd"] Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.183760 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9vvt\" (UniqueName: \"kubernetes.io/projected/ca7d8dda-6139-474a-844e-4a7b182ceb81-kube-api-access-m9vvt\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.183798 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-utilities\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.183901 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-catalog-content\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.287231 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9vvt\" (UniqueName: \"kubernetes.io/projected/ca7d8dda-6139-474a-844e-4a7b182ceb81-kube-api-access-m9vvt\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.287288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-utilities\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.287417 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-catalog-content\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.287948 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-catalog-content\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.288527 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-utilities\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.310449 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9vvt\" (UniqueName: \"kubernetes.io/projected/ca7d8dda-6139-474a-844e-4a7b182ceb81-kube-api-access-m9vvt\") pod \"community-operators-64lbd\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:07 crc kubenswrapper[4696]: I1125 11:32:07.490259 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:08 crc kubenswrapper[4696]: I1125 11:32:08.167968 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-64lbd"] Nov 25 11:32:09 crc kubenswrapper[4696]: I1125 11:32:09.100223 4696 generic.go:334] "Generic (PLEG): container finished" podID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerID="213d43e29f3f2d02447e763fcd535e03dc94ed7a55298a4d942a3d99fac44ae2" exitCode=0 Nov 25 11:32:09 crc kubenswrapper[4696]: I1125 11:32:09.100387 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerDied","Data":"213d43e29f3f2d02447e763fcd535e03dc94ed7a55298a4d942a3d99fac44ae2"} Nov 25 11:32:09 crc kubenswrapper[4696]: I1125 11:32:09.100891 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerStarted","Data":"c9db5d24d3242f415aa1a72f14c14598747e95f1ac335ca27635d84bbfe9abfe"} Nov 25 11:32:13 crc kubenswrapper[4696]: I1125 11:32:13.043585 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:32:13 crc kubenswrapper[4696]: E1125 11:32:13.044426 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:32:26 crc kubenswrapper[4696]: I1125 11:32:26.042983 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:32:26 crc kubenswrapper[4696]: E1125 11:32:26.043567 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:32:26 crc kubenswrapper[4696]: E1125 11:32:26.437603 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 11:32:26 crc kubenswrapper[4696]: E1125 11:32:26.437734 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mkm2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2jlsf_openshift-marketplace(a4329258-c066-423c-b084-db2267dddec8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 11:32:26 crc kubenswrapper[4696]: E1125 11:32:26.439839 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" Nov 25 11:32:27 crc kubenswrapper[4696]: I1125 11:32:27.280028 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hlrx" event={"ID":"25e551df-fa3d-40bd-b2e0-e972976f5c74","Type":"ContainerStarted","Data":"7dbca0913f31ac5cef2308145a3c067f37b794e6716036fce0b501133f9afe6b"} Nov 25 11:32:27 crc kubenswrapper[4696]: I1125 11:32:27.282333 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerStarted","Data":"7848943ef30d6408ae234e35cca4a454132b993363d979b8e40551457b9e3dda"} Nov 25 11:32:27 crc kubenswrapper[4696]: E1125 11:32:27.283827 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" Nov 25 11:32:32 crc kubenswrapper[4696]: I1125 11:32:32.334417 4696 generic.go:334] "Generic (PLEG): container finished" podID="25e551df-fa3d-40bd-b2e0-e972976f5c74" containerID="7dbca0913f31ac5cef2308145a3c067f37b794e6716036fce0b501133f9afe6b" exitCode=0 Nov 25 11:32:32 crc kubenswrapper[4696]: I1125 11:32:32.334498 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hlrx" event={"ID":"25e551df-fa3d-40bd-b2e0-e972976f5c74","Type":"ContainerDied","Data":"7dbca0913f31ac5cef2308145a3c067f37b794e6716036fce0b501133f9afe6b"} Nov 25 11:32:33 crc kubenswrapper[4696]: I1125 11:32:33.346774 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hlrx" event={"ID":"25e551df-fa3d-40bd-b2e0-e972976f5c74","Type":"ContainerStarted","Data":"f9f61ca2091373403dd9d7a68c6fc9e8a78a0dcced39b92b06e7e083692a7a08"} Nov 25 11:32:33 crc kubenswrapper[4696]: I1125 11:32:33.349426 4696 generic.go:334] "Generic (PLEG): container finished" podID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerID="7848943ef30d6408ae234e35cca4a454132b993363d979b8e40551457b9e3dda" exitCode=0 Nov 25 11:32:33 crc kubenswrapper[4696]: I1125 11:32:33.349457 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerDied","Data":"7848943ef30d6408ae234e35cca4a454132b993363d979b8e40551457b9e3dda"} Nov 25 11:32:33 crc kubenswrapper[4696]: I1125 11:32:33.384647 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9hlrx" podStartSLOduration=2.569836894 podStartE2EDuration="28.384627172s" podCreationTimestamp="2025-11-25 11:32:05 +0000 UTC" firstStartedPulling="2025-11-25 11:32:07.072880797 +0000 UTC m=+3323.886497384" lastFinishedPulling="2025-11-25 11:32:32.887671075 +0000 UTC m=+3349.701287662" observedRunningTime="2025-11-25 11:32:33.380261453 +0000 UTC m=+3350.193878040" watchObservedRunningTime="2025-11-25 11:32:33.384627172 +0000 UTC m=+3350.198243759" Nov 25 11:32:34 crc kubenswrapper[4696]: I1125 11:32:34.361803 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerStarted","Data":"ae51e3efbbf029687a557eec25bb7a68b39704169238d0e3ccb7f193f9c62143"} Nov 25 11:32:34 crc kubenswrapper[4696]: I1125 11:32:34.390845 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-64lbd" podStartSLOduration=2.548923731 podStartE2EDuration="27.390827734s" podCreationTimestamp="2025-11-25 11:32:07 +0000 UTC" firstStartedPulling="2025-11-25 11:32:09.102161418 +0000 UTC m=+3325.915778005" lastFinishedPulling="2025-11-25 11:32:33.944065421 +0000 UTC m=+3350.757682008" observedRunningTime="2025-11-25 11:32:34.388969175 +0000 UTC m=+3351.202585772" watchObservedRunningTime="2025-11-25 11:32:34.390827734 +0000 UTC m=+3351.204444321" Nov 25 11:32:35 crc kubenswrapper[4696]: I1125 11:32:35.665710 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:35 crc kubenswrapper[4696]: I1125 11:32:35.666017 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:32:36 crc kubenswrapper[4696]: I1125 11:32:36.716787 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9hlrx" podUID="25e551df-fa3d-40bd-b2e0-e972976f5c74" containerName="registry-server" probeResult="failure" output=< Nov 25 11:32:36 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:32:36 crc kubenswrapper[4696]: > Nov 25 11:32:37 crc kubenswrapper[4696]: I1125 11:32:37.491574 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:37 crc kubenswrapper[4696]: I1125 11:32:37.491653 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:37 crc kubenswrapper[4696]: I1125 11:32:37.554233 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:41 crc kubenswrapper[4696]: I1125 11:32:41.042519 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:32:41 crc kubenswrapper[4696]: E1125 11:32:41.043213 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:32:44 crc kubenswrapper[4696]: I1125 11:32:44.460292 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerStarted","Data":"4b535fe280cb6c88274216f56652de2aae7cdbc3f1533816c022550fb1f9ebe5"} Nov 25 11:32:46 crc kubenswrapper[4696]: I1125 11:32:46.717180 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9hlrx" podUID="25e551df-fa3d-40bd-b2e0-e972976f5c74" containerName="registry-server" probeResult="failure" output=< Nov 25 11:32:46 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:32:46 crc kubenswrapper[4696]: > Nov 25 11:32:47 crc kubenswrapper[4696]: I1125 11:32:47.542004 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:48 crc kubenswrapper[4696]: I1125 11:32:48.383802 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-64lbd"] Nov 25 11:32:48 crc kubenswrapper[4696]: I1125 11:32:48.492043 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-64lbd" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="registry-server" containerID="cri-o://ae51e3efbbf029687a557eec25bb7a68b39704169238d0e3ccb7f193f9c62143" gracePeriod=2 Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.511714 4696 generic.go:334] "Generic (PLEG): container finished" podID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerID="ae51e3efbbf029687a557eec25bb7a68b39704169238d0e3ccb7f193f9c62143" exitCode=0 Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.512629 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerDied","Data":"ae51e3efbbf029687a557eec25bb7a68b39704169238d0e3ccb7f193f9c62143"} Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.709757 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.841834 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-catalog-content\") pod \"ca7d8dda-6139-474a-844e-4a7b182ceb81\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.841936 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9vvt\" (UniqueName: \"kubernetes.io/projected/ca7d8dda-6139-474a-844e-4a7b182ceb81-kube-api-access-m9vvt\") pod \"ca7d8dda-6139-474a-844e-4a7b182ceb81\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.842035 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-utilities\") pod \"ca7d8dda-6139-474a-844e-4a7b182ceb81\" (UID: \"ca7d8dda-6139-474a-844e-4a7b182ceb81\") " Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.842774 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-utilities" (OuterVolumeSpecName: "utilities") pod "ca7d8dda-6139-474a-844e-4a7b182ceb81" (UID: "ca7d8dda-6139-474a-844e-4a7b182ceb81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.869892 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7d8dda-6139-474a-844e-4a7b182ceb81-kube-api-access-m9vvt" (OuterVolumeSpecName: "kube-api-access-m9vvt") pod "ca7d8dda-6139-474a-844e-4a7b182ceb81" (UID: "ca7d8dda-6139-474a-844e-4a7b182ceb81"). InnerVolumeSpecName "kube-api-access-m9vvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.916559 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca7d8dda-6139-474a-844e-4a7b182ceb81" (UID: "ca7d8dda-6139-474a-844e-4a7b182ceb81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.944031 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.944076 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9vvt\" (UniqueName: \"kubernetes.io/projected/ca7d8dda-6139-474a-844e-4a7b182ceb81-kube-api-access-m9vvt\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:49 crc kubenswrapper[4696]: I1125 11:32:49.944089 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7d8dda-6139-474a-844e-4a7b182ceb81-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.523297 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64lbd" Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.524068 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64lbd" event={"ID":"ca7d8dda-6139-474a-844e-4a7b182ceb81","Type":"ContainerDied","Data":"c9db5d24d3242f415aa1a72f14c14598747e95f1ac335ca27635d84bbfe9abfe"} Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.524214 4696 scope.go:117] "RemoveContainer" containerID="ae51e3efbbf029687a557eec25bb7a68b39704169238d0e3ccb7f193f9c62143" Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.527471 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4329258-c066-423c-b084-db2267dddec8" containerID="4b535fe280cb6c88274216f56652de2aae7cdbc3f1533816c022550fb1f9ebe5" exitCode=0 Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.527556 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerDied","Data":"4b535fe280cb6c88274216f56652de2aae7cdbc3f1533816c022550fb1f9ebe5"} Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.533205 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.578550 4696 scope.go:117] "RemoveContainer" containerID="7848943ef30d6408ae234e35cca4a454132b993363d979b8e40551457b9e3dda" Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.598937 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-64lbd"] Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.641725 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-64lbd"] Nov 25 11:32:50 crc kubenswrapper[4696]: I1125 11:32:50.642351 4696 scope.go:117] "RemoveContainer" containerID="213d43e29f3f2d02447e763fcd535e03dc94ed7a55298a4d942a3d99fac44ae2" Nov 25 11:32:52 crc kubenswrapper[4696]: I1125 11:32:52.067685 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" path="/var/lib/kubelet/pods/ca7d8dda-6139-474a-844e-4a7b182ceb81/volumes" Nov 25 11:32:52 crc kubenswrapper[4696]: I1125 11:32:52.547285 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerStarted","Data":"6a3fbd239bc0a33da320e2096db29761194952aa2e0e3bac52bdae2c33a22b63"} Nov 25 11:32:52 crc kubenswrapper[4696]: I1125 11:32:52.571228 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2jlsf" podStartSLOduration=3.214996567 podStartE2EDuration="49.571205441s" podCreationTimestamp="2025-11-25 11:32:03 +0000 UTC" firstStartedPulling="2025-11-25 11:32:05.01627778 +0000 UTC m=+3321.829894367" lastFinishedPulling="2025-11-25 11:32:51.372486654 +0000 UTC m=+3368.186103241" observedRunningTime="2025-11-25 11:32:52.568680391 +0000 UTC m=+3369.382296988" watchObservedRunningTime="2025-11-25 11:32:52.571205441 +0000 UTC m=+3369.384822028" Nov 25 11:32:53 crc kubenswrapper[4696]: I1125 11:32:53.042827 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:32:53 crc kubenswrapper[4696]: E1125 11:32:53.043390 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:32:54 crc kubenswrapper[4696]: I1125 11:32:54.052437 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:54 crc kubenswrapper[4696]: I1125 11:32:54.052473 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:32:55 crc kubenswrapper[4696]: I1125 11:32:55.100289 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" probeResult="failure" output=< Nov 25 11:32:55 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:32:55 crc kubenswrapper[4696]: > Nov 25 11:32:56 crc kubenswrapper[4696]: I1125 11:32:56.721920 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9hlrx" podUID="25e551df-fa3d-40bd-b2e0-e972976f5c74" containerName="registry-server" probeResult="failure" output=< Nov 25 11:32:56 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:32:56 crc kubenswrapper[4696]: > Nov 25 11:33:04 crc kubenswrapper[4696]: I1125 11:33:04.051289 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:33:04 crc kubenswrapper[4696]: E1125 11:33:04.052060 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:33:05 crc kubenswrapper[4696]: I1125 11:33:05.097561 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" probeResult="failure" output=< Nov 25 11:33:05 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:33:05 crc kubenswrapper[4696]: > Nov 25 11:33:05 crc kubenswrapper[4696]: I1125 11:33:05.714993 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:33:05 crc kubenswrapper[4696]: I1125 11:33:05.763182 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9hlrx" Nov 25 11:33:05 crc kubenswrapper[4696]: I1125 11:33:05.829337 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9hlrx"] Nov 25 11:33:05 crc kubenswrapper[4696]: I1125 11:33:05.961834 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:33:05 crc kubenswrapper[4696]: I1125 11:33:05.963088 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dswtf" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="registry-server" containerID="cri-o://3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a" gracePeriod=2 Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.628202 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.680817 4696 generic.go:334] "Generic (PLEG): container finished" podID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerID="3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a" exitCode=0 Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.682064 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dswtf" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.682760 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dswtf" event={"ID":"7dd837c4-e52e-4b41-857f-6dfa1944693d","Type":"ContainerDied","Data":"3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a"} Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.682881 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dswtf" event={"ID":"7dd837c4-e52e-4b41-857f-6dfa1944693d","Type":"ContainerDied","Data":"8b9e41d0e8733f5e7fa83d4d17e0d563f01616c4498ac58d52659ec9d5e6dc97"} Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.682961 4696 scope.go:117] "RemoveContainer" containerID="3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.770695 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc7xm\" (UniqueName: \"kubernetes.io/projected/7dd837c4-e52e-4b41-857f-6dfa1944693d-kube-api-access-dc7xm\") pod \"7dd837c4-e52e-4b41-857f-6dfa1944693d\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.770845 4696 scope.go:117] "RemoveContainer" containerID="0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.771944 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-catalog-content\") pod \"7dd837c4-e52e-4b41-857f-6dfa1944693d\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.771997 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-utilities\") pod \"7dd837c4-e52e-4b41-857f-6dfa1944693d\" (UID: \"7dd837c4-e52e-4b41-857f-6dfa1944693d\") " Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.778386 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-utilities" (OuterVolumeSpecName: "utilities") pod "7dd837c4-e52e-4b41-857f-6dfa1944693d" (UID: "7dd837c4-e52e-4b41-857f-6dfa1944693d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.781972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd837c4-e52e-4b41-857f-6dfa1944693d-kube-api-access-dc7xm" (OuterVolumeSpecName: "kube-api-access-dc7xm") pod "7dd837c4-e52e-4b41-857f-6dfa1944693d" (UID: "7dd837c4-e52e-4b41-857f-6dfa1944693d"). InnerVolumeSpecName "kube-api-access-dc7xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.804251 4696 scope.go:117] "RemoveContainer" containerID="fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.874652 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc7xm\" (UniqueName: \"kubernetes.io/projected/7dd837c4-e52e-4b41-857f-6dfa1944693d-kube-api-access-dc7xm\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.874955 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.896955 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7dd837c4-e52e-4b41-857f-6dfa1944693d" (UID: "7dd837c4-e52e-4b41-857f-6dfa1944693d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.933158 4696 scope.go:117] "RemoveContainer" containerID="3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a" Nov 25 11:33:06 crc kubenswrapper[4696]: E1125 11:33:06.933717 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a\": container with ID starting with 3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a not found: ID does not exist" containerID="3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.933825 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a"} err="failed to get container status \"3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a\": rpc error: code = NotFound desc = could not find container \"3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a\": container with ID starting with 3ed4402d15e179b19987a09658dff4c766823e8869ca9685c9cd058a04fa329a not found: ID does not exist" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.933911 4696 scope.go:117] "RemoveContainer" containerID="0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04" Nov 25 11:33:06 crc kubenswrapper[4696]: E1125 11:33:06.934349 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04\": container with ID starting with 0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04 not found: ID does not exist" containerID="0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.934393 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04"} err="failed to get container status \"0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04\": rpc error: code = NotFound desc = could not find container \"0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04\": container with ID starting with 0174444dda9214386685b4a944bac0c3d691fbc58d9ff017a4fbfa7b897cfd04 not found: ID does not exist" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.934444 4696 scope.go:117] "RemoveContainer" containerID="fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a" Nov 25 11:33:06 crc kubenswrapper[4696]: E1125 11:33:06.934791 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a\": container with ID starting with fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a not found: ID does not exist" containerID="fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.934866 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a"} err="failed to get container status \"fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a\": rpc error: code = NotFound desc = could not find container \"fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a\": container with ID starting with fe77d300cb0da10c2d5ce71008e5b5e5de079f196b31fdbbbd03832dd251846a not found: ID does not exist" Nov 25 11:33:06 crc kubenswrapper[4696]: I1125 11:33:06.977154 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd837c4-e52e-4b41-857f-6dfa1944693d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:07 crc kubenswrapper[4696]: I1125 11:33:07.013619 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:33:07 crc kubenswrapper[4696]: I1125 11:33:07.022920 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dswtf"] Nov 25 11:33:08 crc kubenswrapper[4696]: I1125 11:33:08.083557 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" path="/var/lib/kubelet/pods/7dd837c4-e52e-4b41-857f-6dfa1944693d/volumes" Nov 25 11:33:15 crc kubenswrapper[4696]: I1125 11:33:15.042173 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:33:15 crc kubenswrapper[4696]: E1125 11:33:15.042911 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:33:15 crc kubenswrapper[4696]: I1125 11:33:15.112560 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" probeResult="failure" output=< Nov 25 11:33:15 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:33:15 crc kubenswrapper[4696]: > Nov 25 11:33:25 crc kubenswrapper[4696]: I1125 11:33:25.087217 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" probeResult="failure" output=< Nov 25 11:33:25 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:33:25 crc kubenswrapper[4696]: > Nov 25 11:33:29 crc kubenswrapper[4696]: I1125 11:33:29.042875 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:33:29 crc kubenswrapper[4696]: E1125 11:33:29.043744 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:33:34 crc kubenswrapper[4696]: I1125 11:33:34.105802 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:33:34 crc kubenswrapper[4696]: I1125 11:33:34.160834 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:33:34 crc kubenswrapper[4696]: I1125 11:33:34.227903 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:33:34 crc kubenswrapper[4696]: I1125 11:33:34.353879 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7pxd4"] Nov 25 11:33:34 crc kubenswrapper[4696]: I1125 11:33:34.354862 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7pxd4" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="registry-server" containerID="cri-o://df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd" gracePeriod=2 Nov 25 11:33:35 crc kubenswrapper[4696]: I1125 11:33:35.969029 4696 generic.go:334] "Generic (PLEG): container finished" podID="250727c1-0261-49a2-b93f-86ac9235b443" containerID="df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd" exitCode=0 Nov 25 11:33:35 crc kubenswrapper[4696]: I1125 11:33:35.969635 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerDied","Data":"df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd"} Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.300716 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.466817 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-catalog-content\") pod \"250727c1-0261-49a2-b93f-86ac9235b443\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.467201 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb27q\" (UniqueName: \"kubernetes.io/projected/250727c1-0261-49a2-b93f-86ac9235b443-kube-api-access-rb27q\") pod \"250727c1-0261-49a2-b93f-86ac9235b443\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.467335 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-utilities\") pod \"250727c1-0261-49a2-b93f-86ac9235b443\" (UID: \"250727c1-0261-49a2-b93f-86ac9235b443\") " Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.469002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-utilities" (OuterVolumeSpecName: "utilities") pod "250727c1-0261-49a2-b93f-86ac9235b443" (UID: "250727c1-0261-49a2-b93f-86ac9235b443"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.480678 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250727c1-0261-49a2-b93f-86ac9235b443-kube-api-access-rb27q" (OuterVolumeSpecName: "kube-api-access-rb27q") pod "250727c1-0261-49a2-b93f-86ac9235b443" (UID: "250727c1-0261-49a2-b93f-86ac9235b443"). InnerVolumeSpecName "kube-api-access-rb27q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.569344 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb27q\" (UniqueName: \"kubernetes.io/projected/250727c1-0261-49a2-b93f-86ac9235b443-kube-api-access-rb27q\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.569589 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.576594 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "250727c1-0261-49a2-b93f-86ac9235b443" (UID: "250727c1-0261-49a2-b93f-86ac9235b443"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.671544 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250727c1-0261-49a2-b93f-86ac9235b443-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.981082 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pxd4" event={"ID":"250727c1-0261-49a2-b93f-86ac9235b443","Type":"ContainerDied","Data":"8719bf798fbd5402f3a447f533da3d7cc029ae103ab9efca534d27904fd6dc48"} Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.982409 4696 scope.go:117] "RemoveContainer" containerID="df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd" Nov 25 11:33:36 crc kubenswrapper[4696]: I1125 11:33:36.981178 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pxd4" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.007424 4696 scope.go:117] "RemoveContainer" containerID="df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.016413 4696 scope.go:117] "RemoveContainer" containerID="5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d" Nov 25 11:33:37 crc kubenswrapper[4696]: E1125 11:33:37.016524 4696 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_registry-server_redhat-operators-7pxd4_openshift-marketplace_250727c1-0261-49a2-b93f-86ac9235b443_0 in pod sandbox 8719bf798fbd5402f3a447f533da3d7cc029ae103ab9efca534d27904fd6dc48 from index: no such id: 'df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd'" containerID="df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd" Nov 25 11:33:37 crc kubenswrapper[4696]: E1125 11:33:37.016551 4696 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_registry-server_redhat-operators-7pxd4_openshift-marketplace_250727c1-0261-49a2-b93f-86ac9235b443_0 in pod sandbox 8719bf798fbd5402f3a447f533da3d7cc029ae103ab9efca534d27904fd6dc48 from index: no such id: 'df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd'" containerID="df5d971d11c5a0e5ea0d4c756842025cf834bb65ad3cc39514e6f903aa5478cd" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.016939 4696 scope.go:117] "RemoveContainer" containerID="0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.037372 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7pxd4"] Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.045873 4696 scope.go:117] "RemoveContainer" containerID="0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.059926 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7pxd4"] Nov 25 11:33:37 crc kubenswrapper[4696]: E1125 11:33:37.067219 4696 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_extract-utilities_redhat-operators-7pxd4_openshift-marketplace_250727c1-0261-49a2-b93f-86ac9235b443_0 in pod sandbox 8719bf798fbd5402f3a447f533da3d7cc029ae103ab9efca534d27904fd6dc48 from index: no such id: '0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2'" containerID="0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.067324 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2"} err="rpc error: code = Unknown desc = failed to delete container k8s_extract-utilities_redhat-operators-7pxd4_openshift-marketplace_250727c1-0261-49a2-b93f-86ac9235b443_0 in pod sandbox 8719bf798fbd5402f3a447f533da3d7cc029ae103ab9efca534d27904fd6dc48 from index: no such id: '0330727c067c40b04147cfea8a87c6041223fa6ccf28f953125b38db932dc9a2'" Nov 25 11:33:37 crc kubenswrapper[4696]: I1125 11:33:37.067240 4696 scope.go:117] "RemoveContainer" containerID="5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d" Nov 25 11:33:37 crc kubenswrapper[4696]: E1125 11:33:37.068797 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d\": container with ID starting with 5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d not found: ID does not exist" containerID="5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d" Nov 25 11:33:37 crc kubenswrapper[4696]: E1125 11:33:37.068840 4696 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d\": rpc error: code = NotFound desc = could not find container \"5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d\": container with ID starting with 5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d not found: ID does not exist" containerID="5abc87b18bc7ca0e040abfc4ab67d4c1c34a0802e4d44ab40b86fe2f2ed46b0d" Nov 25 11:33:38 crc kubenswrapper[4696]: I1125 11:33:38.055903 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="250727c1-0261-49a2-b93f-86ac9235b443" path="/var/lib/kubelet/pods/250727c1-0261-49a2-b93f-86ac9235b443/volumes" Nov 25 11:33:40 crc kubenswrapper[4696]: I1125 11:33:40.043216 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:33:40 crc kubenswrapper[4696]: E1125 11:33:40.043776 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:33:52 crc kubenswrapper[4696]: I1125 11:33:52.042326 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:33:52 crc kubenswrapper[4696]: E1125 11:33:52.043262 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:34:06 crc kubenswrapper[4696]: I1125 11:34:06.046400 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:34:07 crc kubenswrapper[4696]: I1125 11:34:07.257940 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"6c4baa78f0aef53fe484d15ca73a565a7a9aae0b3baf5cc43b069fc4fe2095ad"} Nov 25 11:36:30 crc kubenswrapper[4696]: I1125 11:36:30.802080 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:36:30 crc kubenswrapper[4696]: I1125 11:36:30.802646 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:37:00 crc kubenswrapper[4696]: I1125 11:37:00.802417 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:37:00 crc kubenswrapper[4696]: I1125 11:37:00.803076 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:37:30 crc kubenswrapper[4696]: I1125 11:37:30.830188 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:37:30 crc kubenswrapper[4696]: I1125 11:37:30.830776 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:37:30 crc kubenswrapper[4696]: I1125 11:37:30.830825 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:37:30 crc kubenswrapper[4696]: I1125 11:37:30.831643 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c4baa78f0aef53fe484d15ca73a565a7a9aae0b3baf5cc43b069fc4fe2095ad"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:37:30 crc kubenswrapper[4696]: I1125 11:37:30.831726 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://6c4baa78f0aef53fe484d15ca73a565a7a9aae0b3baf5cc43b069fc4fe2095ad" gracePeriod=600 Nov 25 11:37:31 crc kubenswrapper[4696]: I1125 11:37:31.016193 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="6c4baa78f0aef53fe484d15ca73a565a7a9aae0b3baf5cc43b069fc4fe2095ad" exitCode=0 Nov 25 11:37:31 crc kubenswrapper[4696]: I1125 11:37:31.016266 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"6c4baa78f0aef53fe484d15ca73a565a7a9aae0b3baf5cc43b069fc4fe2095ad"} Nov 25 11:37:31 crc kubenswrapper[4696]: I1125 11:37:31.016296 4696 scope.go:117] "RemoveContainer" containerID="f3578b6f36e4029f47c07177e99f9eb3b779911cbc004634383dbdd026129e83" Nov 25 11:37:32 crc kubenswrapper[4696]: I1125 11:37:32.027891 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551"} Nov 25 11:40:00 crc kubenswrapper[4696]: I1125 11:40:00.801914 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:40:00 crc kubenswrapper[4696]: I1125 11:40:00.802463 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:40:30 crc kubenswrapper[4696]: I1125 11:40:30.802512 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:40:30 crc kubenswrapper[4696]: I1125 11:40:30.803617 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:41:00 crc kubenswrapper[4696]: I1125 11:41:00.802023 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:41:00 crc kubenswrapper[4696]: I1125 11:41:00.803588 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:41:00 crc kubenswrapper[4696]: I1125 11:41:00.803977 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:41:00 crc kubenswrapper[4696]: I1125 11:41:00.804881 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:41:00 crc kubenswrapper[4696]: I1125 11:41:00.805043 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" gracePeriod=600 Nov 25 11:41:00 crc kubenswrapper[4696]: E1125 11:41:00.948835 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:41:01 crc kubenswrapper[4696]: I1125 11:41:01.913634 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" exitCode=0 Nov 25 11:41:01 crc kubenswrapper[4696]: I1125 11:41:01.913695 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551"} Nov 25 11:41:01 crc kubenswrapper[4696]: I1125 11:41:01.913764 4696 scope.go:117] "RemoveContainer" containerID="6c4baa78f0aef53fe484d15ca73a565a7a9aae0b3baf5cc43b069fc4fe2095ad" Nov 25 11:41:01 crc kubenswrapper[4696]: I1125 11:41:01.914419 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:41:01 crc kubenswrapper[4696]: E1125 11:41:01.914652 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:41:16 crc kubenswrapper[4696]: I1125 11:41:16.042079 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:41:16 crc kubenswrapper[4696]: E1125 11:41:16.042896 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:41:28 crc kubenswrapper[4696]: I1125 11:41:28.043054 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:41:28 crc kubenswrapper[4696]: E1125 11:41:28.043967 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:41:43 crc kubenswrapper[4696]: I1125 11:41:43.042244 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:41:43 crc kubenswrapper[4696]: E1125 11:41:43.042966 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:41:57 crc kubenswrapper[4696]: I1125 11:41:57.042550 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:41:57 crc kubenswrapper[4696]: E1125 11:41:57.044401 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:42:08 crc kubenswrapper[4696]: I1125 11:42:08.044839 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:42:08 crc kubenswrapper[4696]: E1125 11:42:08.045624 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:42:23 crc kubenswrapper[4696]: I1125 11:42:23.042325 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:42:23 crc kubenswrapper[4696]: E1125 11:42:23.043069 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:42:35 crc kubenswrapper[4696]: I1125 11:42:35.042522 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:42:35 crc kubenswrapper[4696]: E1125 11:42:35.043321 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:42:49 crc kubenswrapper[4696]: I1125 11:42:49.042926 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:42:49 crc kubenswrapper[4696]: E1125 11:42:49.044900 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.532204 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c8ktn"] Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.532921 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="extract-content" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.532953 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="extract-content" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.532967 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="extract-utilities" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.532975 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="extract-utilities" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.532991 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.532998 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.533018 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533026 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.533037 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533044 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.533059 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="extract-content" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533066 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="extract-content" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.533080 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="extract-content" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533087 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="extract-content" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.533097 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="extract-utilities" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533105 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="extract-utilities" Nov 25 11:42:51 crc kubenswrapper[4696]: E1125 11:42:51.533116 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="extract-utilities" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533123 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="extract-utilities" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533366 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd837c4-e52e-4b41-857f-6dfa1944693d" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533388 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="250727c1-0261-49a2-b93f-86ac9235b443" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.533404 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7d8dda-6139-474a-844e-4a7b182ceb81" containerName="registry-server" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.535120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.623835 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8ktn"] Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.678164 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-utilities\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.678321 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w874d\" (UniqueName: \"kubernetes.io/projected/49d71f3f-36a2-4ea4-b794-95ff99ff2598-kube-api-access-w874d\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.678379 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-catalog-content\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.779862 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-utilities\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.779962 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w874d\" (UniqueName: \"kubernetes.io/projected/49d71f3f-36a2-4ea4-b794-95ff99ff2598-kube-api-access-w874d\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.780001 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-catalog-content\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.780375 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-utilities\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.780407 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-catalog-content\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.820971 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w874d\" (UniqueName: \"kubernetes.io/projected/49d71f3f-36a2-4ea4-b794-95ff99ff2598-kube-api-access-w874d\") pod \"redhat-marketplace-c8ktn\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:51 crc kubenswrapper[4696]: I1125 11:42:51.864219 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:42:54 crc kubenswrapper[4696]: I1125 11:42:54.114848 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8ktn"] Nov 25 11:42:55 crc kubenswrapper[4696]: I1125 11:42:55.032064 4696 generic.go:334] "Generic (PLEG): container finished" podID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerID="77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb" exitCode=0 Nov 25 11:42:55 crc kubenswrapper[4696]: I1125 11:42:55.032173 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerDied","Data":"77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb"} Nov 25 11:42:55 crc kubenswrapper[4696]: I1125 11:42:55.032372 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerStarted","Data":"69fa10c1c826850569d4937cc244e94806aeea7f90d426c6b77607c763eb6148"} Nov 25 11:42:55 crc kubenswrapper[4696]: I1125 11:42:55.033933 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:42:57 crc kubenswrapper[4696]: I1125 11:42:57.060282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerStarted","Data":"6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521"} Nov 25 11:42:59 crc kubenswrapper[4696]: I1125 11:42:59.080325 4696 generic.go:334] "Generic (PLEG): container finished" podID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerID="6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521" exitCode=0 Nov 25 11:42:59 crc kubenswrapper[4696]: I1125 11:42:59.080395 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerDied","Data":"6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521"} Nov 25 11:43:01 crc kubenswrapper[4696]: I1125 11:43:01.100225 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerStarted","Data":"a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2"} Nov 25 11:43:01 crc kubenswrapper[4696]: I1125 11:43:01.121356 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c8ktn" podStartSLOduration=5.666618495 podStartE2EDuration="10.121334924s" podCreationTimestamp="2025-11-25 11:42:51 +0000 UTC" firstStartedPulling="2025-11-25 11:42:55.033609391 +0000 UTC m=+3971.847225978" lastFinishedPulling="2025-11-25 11:42:59.48832582 +0000 UTC m=+3976.301942407" observedRunningTime="2025-11-25 11:43:01.117358778 +0000 UTC m=+3977.930975375" watchObservedRunningTime="2025-11-25 11:43:01.121334924 +0000 UTC m=+3977.934951511" Nov 25 11:43:01 crc kubenswrapper[4696]: I1125 11:43:01.864712 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:43:01 crc kubenswrapper[4696]: I1125 11:43:01.865080 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:43:01 crc kubenswrapper[4696]: I1125 11:43:01.919428 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:43:04 crc kubenswrapper[4696]: I1125 11:43:04.056584 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:43:04 crc kubenswrapper[4696]: E1125 11:43:04.057158 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:43:11 crc kubenswrapper[4696]: I1125 11:43:11.916433 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:43:11 crc kubenswrapper[4696]: I1125 11:43:11.975563 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8ktn"] Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.196011 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c8ktn" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="registry-server" containerID="cri-o://a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2" gracePeriod=2 Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.873284 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.917516 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-catalog-content\") pod \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.918004 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w874d\" (UniqueName: \"kubernetes.io/projected/49d71f3f-36a2-4ea4-b794-95ff99ff2598-kube-api-access-w874d\") pod \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.918173 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-utilities\") pod \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\" (UID: \"49d71f3f-36a2-4ea4-b794-95ff99ff2598\") " Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.918948 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-utilities" (OuterVolumeSpecName: "utilities") pod "49d71f3f-36a2-4ea4-b794-95ff99ff2598" (UID: "49d71f3f-36a2-4ea4-b794-95ff99ff2598"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.928948 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d71f3f-36a2-4ea4-b794-95ff99ff2598-kube-api-access-w874d" (OuterVolumeSpecName: "kube-api-access-w874d") pod "49d71f3f-36a2-4ea4-b794-95ff99ff2598" (UID: "49d71f3f-36a2-4ea4-b794-95ff99ff2598"). InnerVolumeSpecName "kube-api-access-w874d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:43:12 crc kubenswrapper[4696]: I1125 11:43:12.941104 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49d71f3f-36a2-4ea4-b794-95ff99ff2598" (UID: "49d71f3f-36a2-4ea4-b794-95ff99ff2598"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.020236 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.020287 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d71f3f-36a2-4ea4-b794-95ff99ff2598-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.020303 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w874d\" (UniqueName: \"kubernetes.io/projected/49d71f3f-36a2-4ea4-b794-95ff99ff2598-kube-api-access-w874d\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.207093 4696 generic.go:334] "Generic (PLEG): container finished" podID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerID="a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2" exitCode=0 Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.207150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerDied","Data":"a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2"} Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.207192 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8ktn" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.207221 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8ktn" event={"ID":"49d71f3f-36a2-4ea4-b794-95ff99ff2598","Type":"ContainerDied","Data":"69fa10c1c826850569d4937cc244e94806aeea7f90d426c6b77607c763eb6148"} Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.207301 4696 scope.go:117] "RemoveContainer" containerID="a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.243147 4696 scope.go:117] "RemoveContainer" containerID="6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.266647 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8ktn"] Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.279801 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8ktn"] Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.296045 4696 scope.go:117] "RemoveContainer" containerID="77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.344639 4696 scope.go:117] "RemoveContainer" containerID="a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2" Nov 25 11:43:13 crc kubenswrapper[4696]: E1125 11:43:13.345530 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2\": container with ID starting with a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2 not found: ID does not exist" containerID="a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.345560 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2"} err="failed to get container status \"a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2\": rpc error: code = NotFound desc = could not find container \"a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2\": container with ID starting with a2b913007f070edfff313e5d9650748b13c564db3ce9ea4b732ee6da56d619f2 not found: ID does not exist" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.345579 4696 scope.go:117] "RemoveContainer" containerID="6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521" Nov 25 11:43:13 crc kubenswrapper[4696]: E1125 11:43:13.345927 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521\": container with ID starting with 6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521 not found: ID does not exist" containerID="6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.345995 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521"} err="failed to get container status \"6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521\": rpc error: code = NotFound desc = could not find container \"6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521\": container with ID starting with 6a787480d694790d6a954113903951d992efd04aa3b11ad0179a96cda1dde521 not found: ID does not exist" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.346054 4696 scope.go:117] "RemoveContainer" containerID="77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb" Nov 25 11:43:13 crc kubenswrapper[4696]: E1125 11:43:13.346621 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb\": container with ID starting with 77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb not found: ID does not exist" containerID="77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb" Nov 25 11:43:13 crc kubenswrapper[4696]: I1125 11:43:13.346694 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb"} err="failed to get container status \"77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb\": rpc error: code = NotFound desc = could not find container \"77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb\": container with ID starting with 77fef12956f95ed0838811a47f16d90e8603a077b8b0ad9085d75698f4d595cb not found: ID does not exist" Nov 25 11:43:14 crc kubenswrapper[4696]: I1125 11:43:14.054721 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" path="/var/lib/kubelet/pods/49d71f3f-36a2-4ea4-b794-95ff99ff2598/volumes" Nov 25 11:43:16 crc kubenswrapper[4696]: I1125 11:43:16.042973 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:43:16 crc kubenswrapper[4696]: E1125 11:43:16.043515 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.612975 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l2n9c"] Nov 25 11:43:18 crc kubenswrapper[4696]: E1125 11:43:18.613755 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="extract-utilities" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.613774 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="extract-utilities" Nov 25 11:43:18 crc kubenswrapper[4696]: E1125 11:43:18.613830 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="extract-content" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.613843 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="extract-content" Nov 25 11:43:18 crc kubenswrapper[4696]: E1125 11:43:18.613858 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="registry-server" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.613865 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="registry-server" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.614087 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d71f3f-36a2-4ea4-b794-95ff99ff2598" containerName="registry-server" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.615624 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.626421 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgfw2\" (UniqueName: \"kubernetes.io/projected/d79b1d06-75b6-4c16-94ca-8e28bece5194-kube-api-access-hgfw2\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.626483 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-catalog-content\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.626609 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-utilities\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.628462 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l2n9c"] Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.728757 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgfw2\" (UniqueName: \"kubernetes.io/projected/d79b1d06-75b6-4c16-94ca-8e28bece5194-kube-api-access-hgfw2\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.728829 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-catalog-content\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.728873 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-utilities\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.729311 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-catalog-content\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.729371 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-utilities\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.827448 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgfw2\" (UniqueName: \"kubernetes.io/projected/d79b1d06-75b6-4c16-94ca-8e28bece5194-kube-api-access-hgfw2\") pod \"certified-operators-l2n9c\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:18 crc kubenswrapper[4696]: I1125 11:43:18.934782 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:19 crc kubenswrapper[4696]: I1125 11:43:19.503760 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l2n9c"] Nov 25 11:43:20 crc kubenswrapper[4696]: I1125 11:43:20.283900 4696 generic.go:334] "Generic (PLEG): container finished" podID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerID="2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2" exitCode=0 Nov 25 11:43:20 crc kubenswrapper[4696]: I1125 11:43:20.283988 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerDied","Data":"2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2"} Nov 25 11:43:20 crc kubenswrapper[4696]: I1125 11:43:20.284193 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerStarted","Data":"548f9c4e449eeccb29dec6ac0cb154131630856f59b04f14481f31187c71206e"} Nov 25 11:43:21 crc kubenswrapper[4696]: I1125 11:43:21.295950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerStarted","Data":"6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07"} Nov 25 11:43:22 crc kubenswrapper[4696]: I1125 11:43:22.305473 4696 generic.go:334] "Generic (PLEG): container finished" podID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerID="6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07" exitCode=0 Nov 25 11:43:22 crc kubenswrapper[4696]: I1125 11:43:22.305560 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerDied","Data":"6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07"} Nov 25 11:43:23 crc kubenswrapper[4696]: I1125 11:43:23.319926 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerStarted","Data":"d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8"} Nov 25 11:43:23 crc kubenswrapper[4696]: I1125 11:43:23.348261 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l2n9c" podStartSLOduration=2.917349498 podStartE2EDuration="5.348239764s" podCreationTimestamp="2025-11-25 11:43:18 +0000 UTC" firstStartedPulling="2025-11-25 11:43:20.286277355 +0000 UTC m=+3997.099893942" lastFinishedPulling="2025-11-25 11:43:22.717167621 +0000 UTC m=+3999.530784208" observedRunningTime="2025-11-25 11:43:23.340324934 +0000 UTC m=+4000.153941531" watchObservedRunningTime="2025-11-25 11:43:23.348239764 +0000 UTC m=+4000.161856351" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.003545 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4bqpn"] Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.006927 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.020699 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bqpn"] Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.189656 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-catalog-content\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.189753 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjs9t\" (UniqueName: \"kubernetes.io/projected/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-kube-api-access-cjs9t\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.189898 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-utilities\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.292131 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-utilities\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.292249 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-catalog-content\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.292283 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjs9t\" (UniqueName: \"kubernetes.io/projected/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-kube-api-access-cjs9t\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.292876 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-catalog-content\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.293882 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-utilities\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.315048 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjs9t\" (UniqueName: \"kubernetes.io/projected/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-kube-api-access-cjs9t\") pod \"redhat-operators-4bqpn\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.325038 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:26 crc kubenswrapper[4696]: I1125 11:43:26.804487 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bqpn"] Nov 25 11:43:26 crc kubenswrapper[4696]: W1125 11:43:26.813952 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29ebbd0a_9f2d_464b_aa73_02f0a62d9857.slice/crio-b7efd7ddbb5ef9e7d1c040b3e49adfc65262e86fde0f203022f203317f96eaad WatchSource:0}: Error finding container b7efd7ddbb5ef9e7d1c040b3e49adfc65262e86fde0f203022f203317f96eaad: Status 404 returned error can't find the container with id b7efd7ddbb5ef9e7d1c040b3e49adfc65262e86fde0f203022f203317f96eaad Nov 25 11:43:27 crc kubenswrapper[4696]: I1125 11:43:27.357358 4696 generic.go:334] "Generic (PLEG): container finished" podID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerID="f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec" exitCode=0 Nov 25 11:43:27 crc kubenswrapper[4696]: I1125 11:43:27.357449 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerDied","Data":"f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec"} Nov 25 11:43:27 crc kubenswrapper[4696]: I1125 11:43:27.357635 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerStarted","Data":"b7efd7ddbb5ef9e7d1c040b3e49adfc65262e86fde0f203022f203317f96eaad"} Nov 25 11:43:28 crc kubenswrapper[4696]: I1125 11:43:28.042624 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:43:28 crc kubenswrapper[4696]: E1125 11:43:28.043027 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:43:28 crc kubenswrapper[4696]: I1125 11:43:28.367811 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerStarted","Data":"49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769"} Nov 25 11:43:28 crc kubenswrapper[4696]: I1125 11:43:28.936278 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:28 crc kubenswrapper[4696]: I1125 11:43:28.936521 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:29 crc kubenswrapper[4696]: I1125 11:43:29.170915 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:29 crc kubenswrapper[4696]: I1125 11:43:29.431371 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:31 crc kubenswrapper[4696]: I1125 11:43:31.195240 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l2n9c"] Nov 25 11:43:32 crc kubenswrapper[4696]: I1125 11:43:32.424467 4696 generic.go:334] "Generic (PLEG): container finished" podID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerID="49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769" exitCode=0 Nov 25 11:43:32 crc kubenswrapper[4696]: I1125 11:43:32.424564 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerDied","Data":"49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769"} Nov 25 11:43:32 crc kubenswrapper[4696]: I1125 11:43:32.424926 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l2n9c" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="registry-server" containerID="cri-o://d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8" gracePeriod=2 Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.038134 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.129910 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-utilities\") pod \"d79b1d06-75b6-4c16-94ca-8e28bece5194\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.130046 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-catalog-content\") pod \"d79b1d06-75b6-4c16-94ca-8e28bece5194\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.130122 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgfw2\" (UniqueName: \"kubernetes.io/projected/d79b1d06-75b6-4c16-94ca-8e28bece5194-kube-api-access-hgfw2\") pod \"d79b1d06-75b6-4c16-94ca-8e28bece5194\" (UID: \"d79b1d06-75b6-4c16-94ca-8e28bece5194\") " Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.130450 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-utilities" (OuterVolumeSpecName: "utilities") pod "d79b1d06-75b6-4c16-94ca-8e28bece5194" (UID: "d79b1d06-75b6-4c16-94ca-8e28bece5194"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.131022 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.155036 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d79b1d06-75b6-4c16-94ca-8e28bece5194-kube-api-access-hgfw2" (OuterVolumeSpecName: "kube-api-access-hgfw2") pod "d79b1d06-75b6-4c16-94ca-8e28bece5194" (UID: "d79b1d06-75b6-4c16-94ca-8e28bece5194"). InnerVolumeSpecName "kube-api-access-hgfw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.180830 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d79b1d06-75b6-4c16-94ca-8e28bece5194" (UID: "d79b1d06-75b6-4c16-94ca-8e28bece5194"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.233583 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d79b1d06-75b6-4c16-94ca-8e28bece5194-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.233624 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgfw2\" (UniqueName: \"kubernetes.io/projected/d79b1d06-75b6-4c16-94ca-8e28bece5194-kube-api-access-hgfw2\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.434782 4696 generic.go:334] "Generic (PLEG): container finished" podID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerID="d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8" exitCode=0 Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.434838 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerDied","Data":"d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8"} Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.434879 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l2n9c" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.435207 4696 scope.go:117] "RemoveContainer" containerID="d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.435186 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l2n9c" event={"ID":"d79b1d06-75b6-4c16-94ca-8e28bece5194","Type":"ContainerDied","Data":"548f9c4e449eeccb29dec6ac0cb154131630856f59b04f14481f31187c71206e"} Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.438688 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerStarted","Data":"625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30"} Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.456447 4696 scope.go:117] "RemoveContainer" containerID="6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.465656 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4bqpn" podStartSLOduration=2.725779137 podStartE2EDuration="8.465637263s" podCreationTimestamp="2025-11-25 11:43:25 +0000 UTC" firstStartedPulling="2025-11-25 11:43:27.360603621 +0000 UTC m=+4004.174220208" lastFinishedPulling="2025-11-25 11:43:33.100461747 +0000 UTC m=+4009.914078334" observedRunningTime="2025-11-25 11:43:33.46338033 +0000 UTC m=+4010.276996927" watchObservedRunningTime="2025-11-25 11:43:33.465637263 +0000 UTC m=+4010.279253850" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.496637 4696 scope.go:117] "RemoveContainer" containerID="2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.512830 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l2n9c"] Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.536123 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l2n9c"] Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.540956 4696 scope.go:117] "RemoveContainer" containerID="d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8" Nov 25 11:43:33 crc kubenswrapper[4696]: E1125 11:43:33.541375 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8\": container with ID starting with d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8 not found: ID does not exist" containerID="d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.541405 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8"} err="failed to get container status \"d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8\": rpc error: code = NotFound desc = could not find container \"d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8\": container with ID starting with d7346f68d4bd810cae8ed4315e825d301cfad3f4d077c8ed719b58c3c5831bd8 not found: ID does not exist" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.541426 4696 scope.go:117] "RemoveContainer" containerID="6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07" Nov 25 11:43:33 crc kubenswrapper[4696]: E1125 11:43:33.542055 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07\": container with ID starting with 6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07 not found: ID does not exist" containerID="6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.542078 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07"} err="failed to get container status \"6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07\": rpc error: code = NotFound desc = could not find container \"6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07\": container with ID starting with 6acf2f4c50de2178a378505d13578f0fde4fb858b030e9026d1463e5263dee07 not found: ID does not exist" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.542091 4696 scope.go:117] "RemoveContainer" containerID="2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2" Nov 25 11:43:33 crc kubenswrapper[4696]: E1125 11:43:33.542336 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2\": container with ID starting with 2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2 not found: ID does not exist" containerID="2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2" Nov 25 11:43:33 crc kubenswrapper[4696]: I1125 11:43:33.542360 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2"} err="failed to get container status \"2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2\": rpc error: code = NotFound desc = could not find container \"2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2\": container with ID starting with 2586e22cda1f497eeb99a97246551f1ab46454a67fddcc42ccd3a6222ec7c2f2 not found: ID does not exist" Nov 25 11:43:34 crc kubenswrapper[4696]: I1125 11:43:34.054150 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" path="/var/lib/kubelet/pods/d79b1d06-75b6-4c16-94ca-8e28bece5194/volumes" Nov 25 11:43:36 crc kubenswrapper[4696]: I1125 11:43:36.325727 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:36 crc kubenswrapper[4696]: I1125 11:43:36.326300 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:37 crc kubenswrapper[4696]: I1125 11:43:37.384653 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4bqpn" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="registry-server" probeResult="failure" output=< Nov 25 11:43:37 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:43:37 crc kubenswrapper[4696]: > Nov 25 11:43:43 crc kubenswrapper[4696]: I1125 11:43:43.043240 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:43:43 crc kubenswrapper[4696]: E1125 11:43:43.044099 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:43:46 crc kubenswrapper[4696]: I1125 11:43:46.371248 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:46 crc kubenswrapper[4696]: I1125 11:43:46.427871 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:46 crc kubenswrapper[4696]: I1125 11:43:46.610442 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bqpn"] Nov 25 11:43:47 crc kubenswrapper[4696]: I1125 11:43:47.571232 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4bqpn" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="registry-server" containerID="cri-o://625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30" gracePeriod=2 Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.181194 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.276578 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-catalog-content\") pod \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.276904 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjs9t\" (UniqueName: \"kubernetes.io/projected/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-kube-api-access-cjs9t\") pod \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.276933 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-utilities\") pod \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\" (UID: \"29ebbd0a-9f2d-464b-aa73-02f0a62d9857\") " Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.278051 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-utilities" (OuterVolumeSpecName: "utilities") pod "29ebbd0a-9f2d-464b-aa73-02f0a62d9857" (UID: "29ebbd0a-9f2d-464b-aa73-02f0a62d9857"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.282751 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-kube-api-access-cjs9t" (OuterVolumeSpecName: "kube-api-access-cjs9t") pod "29ebbd0a-9f2d-464b-aa73-02f0a62d9857" (UID: "29ebbd0a-9f2d-464b-aa73-02f0a62d9857"). InnerVolumeSpecName "kube-api-access-cjs9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.380732 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjs9t\" (UniqueName: \"kubernetes.io/projected/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-kube-api-access-cjs9t\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.381019 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.393385 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29ebbd0a-9f2d-464b-aa73-02f0a62d9857" (UID: "29ebbd0a-9f2d-464b-aa73-02f0a62d9857"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.482643 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ebbd0a-9f2d-464b-aa73-02f0a62d9857-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.581089 4696 generic.go:334] "Generic (PLEG): container finished" podID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerID="625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30" exitCode=0 Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.581126 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bqpn" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.581137 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerDied","Data":"625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30"} Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.581169 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bqpn" event={"ID":"29ebbd0a-9f2d-464b-aa73-02f0a62d9857","Type":"ContainerDied","Data":"b7efd7ddbb5ef9e7d1c040b3e49adfc65262e86fde0f203022f203317f96eaad"} Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.581191 4696 scope.go:117] "RemoveContainer" containerID="625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.604046 4696 scope.go:117] "RemoveContainer" containerID="49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769" Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.618065 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bqpn"] Nov 25 11:43:48 crc kubenswrapper[4696]: I1125 11:43:48.629230 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4bqpn"] Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.040117 4696 scope.go:117] "RemoveContainer" containerID="f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec" Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.079758 4696 scope.go:117] "RemoveContainer" containerID="625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30" Nov 25 11:43:49 crc kubenswrapper[4696]: E1125 11:43:49.080315 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30\": container with ID starting with 625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30 not found: ID does not exist" containerID="625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30" Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.080389 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30"} err="failed to get container status \"625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30\": rpc error: code = NotFound desc = could not find container \"625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30\": container with ID starting with 625a8e19efcad1643f7fb2c75a53b8c25f39521627a8196bacf70c461bf0bd30 not found: ID does not exist" Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.080422 4696 scope.go:117] "RemoveContainer" containerID="49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769" Nov 25 11:43:49 crc kubenswrapper[4696]: E1125 11:43:49.080867 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769\": container with ID starting with 49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769 not found: ID does not exist" containerID="49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769" Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.080898 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769"} err="failed to get container status \"49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769\": rpc error: code = NotFound desc = could not find container \"49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769\": container with ID starting with 49b5fc4ac2f5ad4442536ef88e299d74d90feff0538c39a864965c6af5bba769 not found: ID does not exist" Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.080916 4696 scope.go:117] "RemoveContainer" containerID="f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec" Nov 25 11:43:49 crc kubenswrapper[4696]: E1125 11:43:49.081278 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec\": container with ID starting with f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec not found: ID does not exist" containerID="f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec" Nov 25 11:43:49 crc kubenswrapper[4696]: I1125 11:43:49.081307 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec"} err="failed to get container status \"f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec\": rpc error: code = NotFound desc = could not find container \"f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec\": container with ID starting with f84ae0eef1cab0883b10cbf64657320e60239477b1d1305477d8b3d1b3cf9cec not found: ID does not exist" Nov 25 11:43:50 crc kubenswrapper[4696]: I1125 11:43:50.053753 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" path="/var/lib/kubelet/pods/29ebbd0a-9f2d-464b-aa73-02f0a62d9857/volumes" Nov 25 11:43:54 crc kubenswrapper[4696]: I1125 11:43:54.048504 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:43:54 crc kubenswrapper[4696]: E1125 11:43:54.049454 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.117864 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ndrjf"] Nov 25 11:44:01 crc kubenswrapper[4696]: E1125 11:44:01.118797 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="extract-utilities" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.118814 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="extract-utilities" Nov 25 11:44:01 crc kubenswrapper[4696]: E1125 11:44:01.118840 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="extract-content" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.118848 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="extract-content" Nov 25 11:44:01 crc kubenswrapper[4696]: E1125 11:44:01.118865 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="registry-server" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.118873 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="registry-server" Nov 25 11:44:01 crc kubenswrapper[4696]: E1125 11:44:01.118882 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="extract-utilities" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.118890 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="extract-utilities" Nov 25 11:44:01 crc kubenswrapper[4696]: E1125 11:44:01.118918 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="extract-content" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.118926 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="extract-content" Nov 25 11:44:01 crc kubenswrapper[4696]: E1125 11:44:01.118946 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="registry-server" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.118954 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="registry-server" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.119188 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79b1d06-75b6-4c16-94ca-8e28bece5194" containerName="registry-server" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.119227 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ebbd0a-9f2d-464b-aa73-02f0a62d9857" containerName="registry-server" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.120820 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.143655 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndrjf"] Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.233268 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-utilities\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.234723 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2f86\" (UniqueName: \"kubernetes.io/projected/126fff51-fcd7-4e57-8f91-0858a4d16481-kube-api-access-r2f86\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.234832 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-catalog-content\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.337077 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-utilities\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.337171 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2f86\" (UniqueName: \"kubernetes.io/projected/126fff51-fcd7-4e57-8f91-0858a4d16481-kube-api-access-r2f86\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.337204 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-catalog-content\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.337886 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-catalog-content\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.338018 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-utilities\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.360059 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2f86\" (UniqueName: \"kubernetes.io/projected/126fff51-fcd7-4e57-8f91-0858a4d16481-kube-api-access-r2f86\") pod \"community-operators-ndrjf\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:01 crc kubenswrapper[4696]: I1125 11:44:01.464252 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:02 crc kubenswrapper[4696]: I1125 11:44:02.235196 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndrjf"] Nov 25 11:44:02 crc kubenswrapper[4696]: I1125 11:44:02.718556 4696 generic.go:334] "Generic (PLEG): container finished" podID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerID="81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7" exitCode=0 Nov 25 11:44:02 crc kubenswrapper[4696]: I1125 11:44:02.718800 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerDied","Data":"81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7"} Nov 25 11:44:02 crc kubenswrapper[4696]: I1125 11:44:02.718870 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerStarted","Data":"72e06a82ccba37d52ac04521bf27c55cf7558ca2b34cca7187494521788cbf89"} Nov 25 11:44:03 crc kubenswrapper[4696]: I1125 11:44:03.728254 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerStarted","Data":"72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae"} Nov 25 11:44:05 crc kubenswrapper[4696]: I1125 11:44:05.744722 4696 generic.go:334] "Generic (PLEG): container finished" podID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerID="72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae" exitCode=0 Nov 25 11:44:05 crc kubenswrapper[4696]: I1125 11:44:05.744888 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerDied","Data":"72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae"} Nov 25 11:44:06 crc kubenswrapper[4696]: I1125 11:44:06.756579 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerStarted","Data":"ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e"} Nov 25 11:44:06 crc kubenswrapper[4696]: I1125 11:44:06.786418 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ndrjf" podStartSLOduration=2.405788262 podStartE2EDuration="5.786392632s" podCreationTimestamp="2025-11-25 11:44:01 +0000 UTC" firstStartedPulling="2025-11-25 11:44:02.72050819 +0000 UTC m=+4039.534124777" lastFinishedPulling="2025-11-25 11:44:06.10111256 +0000 UTC m=+4042.914729147" observedRunningTime="2025-11-25 11:44:06.77683062 +0000 UTC m=+4043.590447207" watchObservedRunningTime="2025-11-25 11:44:06.786392632 +0000 UTC m=+4043.600009219" Nov 25 11:44:07 crc kubenswrapper[4696]: I1125 11:44:07.042472 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:44:07 crc kubenswrapper[4696]: E1125 11:44:07.042787 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:44:11 crc kubenswrapper[4696]: I1125 11:44:11.465828 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:11 crc kubenswrapper[4696]: I1125 11:44:11.466397 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:11 crc kubenswrapper[4696]: I1125 11:44:11.517648 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:11 crc kubenswrapper[4696]: I1125 11:44:11.844371 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:11 crc kubenswrapper[4696]: I1125 11:44:11.893954 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ndrjf"] Nov 25 11:44:13 crc kubenswrapper[4696]: I1125 11:44:13.812635 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ndrjf" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="registry-server" containerID="cri-o://ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e" gracePeriod=2 Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.421555 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.549902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-catalog-content\") pod \"126fff51-fcd7-4e57-8f91-0858a4d16481\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.550066 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2f86\" (UniqueName: \"kubernetes.io/projected/126fff51-fcd7-4e57-8f91-0858a4d16481-kube-api-access-r2f86\") pod \"126fff51-fcd7-4e57-8f91-0858a4d16481\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.550095 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-utilities\") pod \"126fff51-fcd7-4e57-8f91-0858a4d16481\" (UID: \"126fff51-fcd7-4e57-8f91-0858a4d16481\") " Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.551070 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-utilities" (OuterVolumeSpecName: "utilities") pod "126fff51-fcd7-4e57-8f91-0858a4d16481" (UID: "126fff51-fcd7-4e57-8f91-0858a4d16481"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.556597 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/126fff51-fcd7-4e57-8f91-0858a4d16481-kube-api-access-r2f86" (OuterVolumeSpecName: "kube-api-access-r2f86") pod "126fff51-fcd7-4e57-8f91-0858a4d16481" (UID: "126fff51-fcd7-4e57-8f91-0858a4d16481"). InnerVolumeSpecName "kube-api-access-r2f86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.626688 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "126fff51-fcd7-4e57-8f91-0858a4d16481" (UID: "126fff51-fcd7-4e57-8f91-0858a4d16481"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.652962 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.653011 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2f86\" (UniqueName: \"kubernetes.io/projected/126fff51-fcd7-4e57-8f91-0858a4d16481-kube-api-access-r2f86\") on node \"crc\" DevicePath \"\"" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.653025 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/126fff51-fcd7-4e57-8f91-0858a4d16481-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.830808 4696 generic.go:334] "Generic (PLEG): container finished" podID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerID="ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e" exitCode=0 Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.830863 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerDied","Data":"ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e"} Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.830897 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndrjf" event={"ID":"126fff51-fcd7-4e57-8f91-0858a4d16481","Type":"ContainerDied","Data":"72e06a82ccba37d52ac04521bf27c55cf7558ca2b34cca7187494521788cbf89"} Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.830920 4696 scope.go:117] "RemoveContainer" containerID="ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.831019 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndrjf" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.890620 4696 scope.go:117] "RemoveContainer" containerID="72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae" Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.891498 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ndrjf"] Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.916994 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ndrjf"] Nov 25 11:44:14 crc kubenswrapper[4696]: I1125 11:44:14.951697 4696 scope.go:117] "RemoveContainer" containerID="81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7" Nov 25 11:44:15 crc kubenswrapper[4696]: I1125 11:44:15.013307 4696 scope.go:117] "RemoveContainer" containerID="ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e" Nov 25 11:44:15 crc kubenswrapper[4696]: E1125 11:44:15.016840 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e\": container with ID starting with ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e not found: ID does not exist" containerID="ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e" Nov 25 11:44:15 crc kubenswrapper[4696]: I1125 11:44:15.016882 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e"} err="failed to get container status \"ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e\": rpc error: code = NotFound desc = could not find container \"ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e\": container with ID starting with ed115020bd27da2d2c3e60340b8ae636b91e158ee76dae6877699c76e1ea536e not found: ID does not exist" Nov 25 11:44:15 crc kubenswrapper[4696]: I1125 11:44:15.016913 4696 scope.go:117] "RemoveContainer" containerID="72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae" Nov 25 11:44:15 crc kubenswrapper[4696]: E1125 11:44:15.020754 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae\": container with ID starting with 72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae not found: ID does not exist" containerID="72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae" Nov 25 11:44:15 crc kubenswrapper[4696]: I1125 11:44:15.020786 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae"} err="failed to get container status \"72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae\": rpc error: code = NotFound desc = could not find container \"72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae\": container with ID starting with 72fc4e982e9f5ef4ccfb469863313fd6c46ea491cae8acb0c2764a1fbb9a20ae not found: ID does not exist" Nov 25 11:44:15 crc kubenswrapper[4696]: I1125 11:44:15.020804 4696 scope.go:117] "RemoveContainer" containerID="81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7" Nov 25 11:44:15 crc kubenswrapper[4696]: E1125 11:44:15.025787 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7\": container with ID starting with 81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7 not found: ID does not exist" containerID="81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7" Nov 25 11:44:15 crc kubenswrapper[4696]: I1125 11:44:15.025822 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7"} err="failed to get container status \"81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7\": rpc error: code = NotFound desc = could not find container \"81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7\": container with ID starting with 81d25d4a562e641803719a233f9b90b55d6b177d0d36cdc90c3729cdf8e8dbe7 not found: ID does not exist" Nov 25 11:44:16 crc kubenswrapper[4696]: I1125 11:44:16.058779 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" path="/var/lib/kubelet/pods/126fff51-fcd7-4e57-8f91-0858a4d16481/volumes" Nov 25 11:44:21 crc kubenswrapper[4696]: I1125 11:44:21.043751 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:44:21 crc kubenswrapper[4696]: E1125 11:44:21.045219 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:44:32 crc kubenswrapper[4696]: I1125 11:44:32.042484 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:44:32 crc kubenswrapper[4696]: E1125 11:44:32.043289 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:44:43 crc kubenswrapper[4696]: I1125 11:44:43.042378 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:44:43 crc kubenswrapper[4696]: E1125 11:44:43.043116 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:44:55 crc kubenswrapper[4696]: I1125 11:44:55.042555 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:44:55 crc kubenswrapper[4696]: E1125 11:44:55.043929 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.178868 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm"] Nov 25 11:45:00 crc kubenswrapper[4696]: E1125 11:45:00.179941 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="extract-utilities" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.179957 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="extract-utilities" Nov 25 11:45:00 crc kubenswrapper[4696]: E1125 11:45:00.179984 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="registry-server" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.179992 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="registry-server" Nov 25 11:45:00 crc kubenswrapper[4696]: E1125 11:45:00.180020 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="extract-content" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.180028 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="extract-content" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.180211 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="126fff51-fcd7-4e57-8f91-0858a4d16481" containerName="registry-server" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.180882 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.184866 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.184900 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.191844 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm"] Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.350517 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnhvh\" (UniqueName: \"kubernetes.io/projected/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-kube-api-access-nnhvh\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.350598 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-config-volume\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.350641 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-secret-volume\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.452560 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-config-volume\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.452683 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-secret-volume\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.452939 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnhvh\" (UniqueName: \"kubernetes.io/projected/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-kube-api-access-nnhvh\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.454423 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-config-volume\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.461559 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-secret-volume\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.471757 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnhvh\" (UniqueName: \"kubernetes.io/projected/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-kube-api-access-nnhvh\") pod \"collect-profiles-29401185-lmmlm\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:00 crc kubenswrapper[4696]: I1125 11:45:00.508286 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:01 crc kubenswrapper[4696]: I1125 11:45:01.008308 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm"] Nov 25 11:45:01 crc kubenswrapper[4696]: I1125 11:45:01.277684 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" event={"ID":"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6","Type":"ContainerStarted","Data":"75e9e3bf70ec5dee48d45e5c4c7d55be5ac9bff93d9670c4c5a3f29c5ae9a881"} Nov 25 11:45:01 crc kubenswrapper[4696]: I1125 11:45:01.277976 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" event={"ID":"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6","Type":"ContainerStarted","Data":"cd5112626cd26247f4fda44a72cd807d8fa1c36f5100b130a5aedd237eb6a933"} Nov 25 11:45:01 crc kubenswrapper[4696]: I1125 11:45:01.301696 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" podStartSLOduration=1.301673598 podStartE2EDuration="1.301673598s" podCreationTimestamp="2025-11-25 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:45:01.292345212 +0000 UTC m=+4098.105961819" watchObservedRunningTime="2025-11-25 11:45:01.301673598 +0000 UTC m=+4098.115290185" Nov 25 11:45:02 crc kubenswrapper[4696]: I1125 11:45:02.290082 4696 generic.go:334] "Generic (PLEG): container finished" podID="c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" containerID="75e9e3bf70ec5dee48d45e5c4c7d55be5ac9bff93d9670c4c5a3f29c5ae9a881" exitCode=0 Nov 25 11:45:02 crc kubenswrapper[4696]: I1125 11:45:02.291258 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" event={"ID":"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6","Type":"ContainerDied","Data":"75e9e3bf70ec5dee48d45e5c4c7d55be5ac9bff93d9670c4c5a3f29c5ae9a881"} Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.718045 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.819909 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-config-volume\") pod \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.820075 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-secret-volume\") pod \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.820145 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnhvh\" (UniqueName: \"kubernetes.io/projected/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-kube-api-access-nnhvh\") pod \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\" (UID: \"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6\") " Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.821683 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-config-volume" (OuterVolumeSpecName: "config-volume") pod "c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" (UID: "c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.827004 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" (UID: "c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.827228 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-kube-api-access-nnhvh" (OuterVolumeSpecName: "kube-api-access-nnhvh") pod "c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" (UID: "c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6"). InnerVolumeSpecName "kube-api-access-nnhvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.922132 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.922349 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:03 crc kubenswrapper[4696]: I1125 11:45:03.922405 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnhvh\" (UniqueName: \"kubernetes.io/projected/c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6-kube-api-access-nnhvh\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:04 crc kubenswrapper[4696]: I1125 11:45:04.318283 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" event={"ID":"c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6","Type":"ContainerDied","Data":"cd5112626cd26247f4fda44a72cd807d8fa1c36f5100b130a5aedd237eb6a933"} Nov 25 11:45:04 crc kubenswrapper[4696]: I1125 11:45:04.318993 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd5112626cd26247f4fda44a72cd807d8fa1c36f5100b130a5aedd237eb6a933" Nov 25 11:45:04 crc kubenswrapper[4696]: I1125 11:45:04.318966 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-lmmlm" Nov 25 11:45:04 crc kubenswrapper[4696]: I1125 11:45:04.370470 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl"] Nov 25 11:45:04 crc kubenswrapper[4696]: I1125 11:45:04.378776 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-56scl"] Nov 25 11:45:06 crc kubenswrapper[4696]: I1125 11:45:06.043417 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:45:06 crc kubenswrapper[4696]: E1125 11:45:06.043987 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:45:06 crc kubenswrapper[4696]: I1125 11:45:06.065821 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="510e1af6-4069-48d0-a02c-17799da5eeed" path="/var/lib/kubelet/pods/510e1af6-4069-48d0-a02c-17799da5eeed/volumes" Nov 25 11:45:18 crc kubenswrapper[4696]: I1125 11:45:18.043632 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:45:18 crc kubenswrapper[4696]: E1125 11:45:18.044490 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:45:32 crc kubenswrapper[4696]: I1125 11:45:32.042965 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:45:32 crc kubenswrapper[4696]: E1125 11:45:32.043784 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:45:37 crc kubenswrapper[4696]: I1125 11:45:37.433225 4696 scope.go:117] "RemoveContainer" containerID="26fdd17402683ff0fb6db77b0b89ff28f50fa2234474262337406fe22d85039a" Nov 25 11:45:45 crc kubenswrapper[4696]: I1125 11:45:45.042840 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:45:45 crc kubenswrapper[4696]: E1125 11:45:45.043699 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:45:57 crc kubenswrapper[4696]: I1125 11:45:57.042372 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:45:57 crc kubenswrapper[4696]: E1125 11:45:57.043220 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:46:12 crc kubenswrapper[4696]: I1125 11:46:12.045974 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:46:12 crc kubenswrapper[4696]: I1125 11:46:12.981522 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"0b65c8a742c0d66f761404eda52c203cf3b833d07f3569d491d2bdd59d6f3b40"} Nov 25 11:48:30 crc kubenswrapper[4696]: I1125 11:48:30.802720 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:48:30 crc kubenswrapper[4696]: I1125 11:48:30.804396 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:49:00 crc kubenswrapper[4696]: I1125 11:49:00.801593 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:49:00 crc kubenswrapper[4696]: I1125 11:49:00.803583 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:49:30 crc kubenswrapper[4696]: I1125 11:49:30.802033 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:49:30 crc kubenswrapper[4696]: I1125 11:49:30.802717 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:49:30 crc kubenswrapper[4696]: I1125 11:49:30.802776 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:49:30 crc kubenswrapper[4696]: I1125 11:49:30.803617 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b65c8a742c0d66f761404eda52c203cf3b833d07f3569d491d2bdd59d6f3b40"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:49:30 crc kubenswrapper[4696]: I1125 11:49:30.803709 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://0b65c8a742c0d66f761404eda52c203cf3b833d07f3569d491d2bdd59d6f3b40" gracePeriod=600 Nov 25 11:49:31 crc kubenswrapper[4696]: I1125 11:49:31.821718 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="0b65c8a742c0d66f761404eda52c203cf3b833d07f3569d491d2bdd59d6f3b40" exitCode=0 Nov 25 11:49:31 crc kubenswrapper[4696]: I1125 11:49:31.821834 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"0b65c8a742c0d66f761404eda52c203cf3b833d07f3569d491d2bdd59d6f3b40"} Nov 25 11:49:31 crc kubenswrapper[4696]: I1125 11:49:31.822314 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b"} Nov 25 11:49:31 crc kubenswrapper[4696]: I1125 11:49:31.822336 4696 scope.go:117] "RemoveContainer" containerID="1726bafcdd61e24b2f6a9305f04a8362e98ef71dfb27bb1be7ecad3a509b5551" Nov 25 11:52:00 crc kubenswrapper[4696]: I1125 11:52:00.802501 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:52:00 crc kubenswrapper[4696]: I1125 11:52:00.803029 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:52:09 crc kubenswrapper[4696]: I1125 11:52:09.198554 4696 generic.go:334] "Generic (PLEG): container finished" podID="f8d03f21-fad7-432e-8220-12094b55a1a1" containerID="7a2a28c3619519f7d9e3fc6ad12511859129535bb6d30665e6fd9690f9995610" exitCode=0 Nov 25 11:52:09 crc kubenswrapper[4696]: I1125 11:52:09.198707 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f8d03f21-fad7-432e-8220-12094b55a1a1","Type":"ContainerDied","Data":"7a2a28c3619519f7d9e3fc6ad12511859129535bb6d30665e6fd9690f9995610"} Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.667558 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806277 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-config-data\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806330 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config-secret\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806382 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5tzh\" (UniqueName: \"kubernetes.io/projected/f8d03f21-fad7-432e-8220-12094b55a1a1-kube-api-access-d5tzh\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806413 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-workdir\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806451 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806546 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ssh-key\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806572 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-temporary\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806627 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ca-certs\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.806729 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f8d03f21-fad7-432e-8220-12094b55a1a1\" (UID: \"f8d03f21-fad7-432e-8220-12094b55a1a1\") " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.808876 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.809428 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-config-data" (OuterVolumeSpecName: "config-data") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.815593 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d03f21-fad7-432e-8220-12094b55a1a1-kube-api-access-d5tzh" (OuterVolumeSpecName: "kube-api-access-d5tzh") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "kube-api-access-d5tzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.817489 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.820058 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.845979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.852434 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.864592 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.878039 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f8d03f21-fad7-432e-8220-12094b55a1a1" (UID: "f8d03f21-fad7-432e-8220-12094b55a1a1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.915794 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.915850 4696 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.915867 4696 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.916556 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.916586 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.916603 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.916617 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5tzh\" (UniqueName: \"kubernetes.io/projected/f8d03f21-fad7-432e-8220-12094b55a1a1-kube-api-access-d5tzh\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.916629 4696 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f8d03f21-fad7-432e-8220-12094b55a1a1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.916640 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8d03f21-fad7-432e-8220-12094b55a1a1-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:10 crc kubenswrapper[4696]: I1125 11:52:10.946587 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 11:52:11 crc kubenswrapper[4696]: I1125 11:52:11.019564 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:11 crc kubenswrapper[4696]: I1125 11:52:11.216320 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f8d03f21-fad7-432e-8220-12094b55a1a1","Type":"ContainerDied","Data":"3e59e4a28d02156283086c974188223877485ed0006fe7d6e50cdf162a345764"} Nov 25 11:52:11 crc kubenswrapper[4696]: I1125 11:52:11.216359 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e59e4a28d02156283086c974188223877485ed0006fe7d6e50cdf162a345764" Nov 25 11:52:11 crc kubenswrapper[4696]: I1125 11:52:11.216414 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.001693 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 11:52:15 crc kubenswrapper[4696]: E1125 11:52:15.002487 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" containerName="collect-profiles" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.002499 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" containerName="collect-profiles" Nov 25 11:52:15 crc kubenswrapper[4696]: E1125 11:52:15.002522 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d03f21-fad7-432e-8220-12094b55a1a1" containerName="tempest-tests-tempest-tests-runner" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.002529 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d03f21-fad7-432e-8220-12094b55a1a1" containerName="tempest-tests-tempest-tests-runner" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.002697 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f9545b-959a-4f2a-9f3c-0e4e42d23bf6" containerName="collect-profiles" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.002720 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d03f21-fad7-432e-8220-12094b55a1a1" containerName="tempest-tests-tempest-tests-runner" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.003293 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.006143 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-58qxz" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.017013 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.097251 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgjwh\" (UniqueName: \"kubernetes.io/projected/5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6-kube-api-access-sgjwh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.097708 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.200249 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.200435 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgjwh\" (UniqueName: \"kubernetes.io/projected/5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6-kube-api-access-sgjwh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.203133 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.261357 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgjwh\" (UniqueName: \"kubernetes.io/projected/5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6-kube-api-access-sgjwh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.262153 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:15 crc kubenswrapper[4696]: I1125 11:52:15.337557 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:52:16 crc kubenswrapper[4696]: I1125 11:52:16.230303 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 11:52:16 crc kubenswrapper[4696]: I1125 11:52:16.234538 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:52:16 crc kubenswrapper[4696]: I1125 11:52:16.296613 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6","Type":"ContainerStarted","Data":"1d687bbabab2299358e10be3d244e993bc598720229223e649bb8868ac1389b8"} Nov 25 11:52:18 crc kubenswrapper[4696]: I1125 11:52:18.315986 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6","Type":"ContainerStarted","Data":"50336fd75aa794477f9f229d191a2b81dcb32719fcebdf7661ace0e5c012ff3c"} Nov 25 11:52:18 crc kubenswrapper[4696]: I1125 11:52:18.336133 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.406603792 podStartE2EDuration="4.336108338s" podCreationTimestamp="2025-11-25 11:52:14 +0000 UTC" firstStartedPulling="2025-11-25 11:52:16.23430001 +0000 UTC m=+4533.047916597" lastFinishedPulling="2025-11-25 11:52:17.163804556 +0000 UTC m=+4533.977421143" observedRunningTime="2025-11-25 11:52:18.329337844 +0000 UTC m=+4535.142954431" watchObservedRunningTime="2025-11-25 11:52:18.336108338 +0000 UTC m=+4535.149724925" Nov 25 11:52:30 crc kubenswrapper[4696]: I1125 11:52:30.801919 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:52:30 crc kubenswrapper[4696]: I1125 11:52:30.802459 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.575890 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-szjd4/must-gather-lg95d"] Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.578029 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.580330 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-szjd4"/"default-dockercfg-v4t8d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.583363 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-szjd4"/"openshift-service-ca.crt" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.585619 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-szjd4"/"kube-root-ca.crt" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.604344 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-szjd4/must-gather-lg95d"] Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.662247 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-must-gather-output\") pod \"must-gather-lg95d\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.662437 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvp7v\" (UniqueName: \"kubernetes.io/projected/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-kube-api-access-hvp7v\") pod \"must-gather-lg95d\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.764399 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvp7v\" (UniqueName: \"kubernetes.io/projected/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-kube-api-access-hvp7v\") pod \"must-gather-lg95d\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.764901 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-must-gather-output\") pod \"must-gather-lg95d\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.765303 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-must-gather-output\") pod \"must-gather-lg95d\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.788588 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvp7v\" (UniqueName: \"kubernetes.io/projected/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-kube-api-access-hvp7v\") pod \"must-gather-lg95d\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:42 crc kubenswrapper[4696]: I1125 11:52:42.901151 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:52:43 crc kubenswrapper[4696]: I1125 11:52:43.442101 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-szjd4/must-gather-lg95d"] Nov 25 11:52:44 crc kubenswrapper[4696]: I1125 11:52:44.549721 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/must-gather-lg95d" event={"ID":"60f42251-7fe6-4fdb-aaec-fbafe9c9647c","Type":"ContainerStarted","Data":"743d2bf8a563f00c0b55fe06a27b66c846d502e0a9b47323db4827d8cc62ca20"} Nov 25 11:52:49 crc kubenswrapper[4696]: I1125 11:52:49.603555 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/must-gather-lg95d" event={"ID":"60f42251-7fe6-4fdb-aaec-fbafe9c9647c","Type":"ContainerStarted","Data":"f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4"} Nov 25 11:52:49 crc kubenswrapper[4696]: I1125 11:52:49.604188 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/must-gather-lg95d" event={"ID":"60f42251-7fe6-4fdb-aaec-fbafe9c9647c","Type":"ContainerStarted","Data":"2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0"} Nov 25 11:52:49 crc kubenswrapper[4696]: I1125 11:52:49.621863 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-szjd4/must-gather-lg95d" podStartSLOduration=2.62390197 podStartE2EDuration="7.621842242s" podCreationTimestamp="2025-11-25 11:52:42 +0000 UTC" firstStartedPulling="2025-11-25 11:52:43.83293396 +0000 UTC m=+4560.646550547" lastFinishedPulling="2025-11-25 11:52:48.830874232 +0000 UTC m=+4565.644490819" observedRunningTime="2025-11-25 11:52:49.618816096 +0000 UTC m=+4566.432432673" watchObservedRunningTime="2025-11-25 11:52:49.621842242 +0000 UTC m=+4566.435458829" Nov 25 11:52:56 crc kubenswrapper[4696]: I1125 11:52:56.846390 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-szjd4/crc-debug-98lgp"] Nov 25 11:52:56 crc kubenswrapper[4696]: I1125 11:52:56.849076 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:56 crc kubenswrapper[4696]: I1125 11:52:56.958904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b178e2c-b39e-4b45-afec-c81005eddebc-host\") pod \"crc-debug-98lgp\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:56 crc kubenswrapper[4696]: I1125 11:52:56.958958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6kf6\" (UniqueName: \"kubernetes.io/projected/4b178e2c-b39e-4b45-afec-c81005eddebc-kube-api-access-r6kf6\") pod \"crc-debug-98lgp\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:57 crc kubenswrapper[4696]: I1125 11:52:57.060724 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b178e2c-b39e-4b45-afec-c81005eddebc-host\") pod \"crc-debug-98lgp\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:57 crc kubenswrapper[4696]: I1125 11:52:57.061135 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6kf6\" (UniqueName: \"kubernetes.io/projected/4b178e2c-b39e-4b45-afec-c81005eddebc-kube-api-access-r6kf6\") pod \"crc-debug-98lgp\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:57 crc kubenswrapper[4696]: I1125 11:52:57.060875 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b178e2c-b39e-4b45-afec-c81005eddebc-host\") pod \"crc-debug-98lgp\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:57 crc kubenswrapper[4696]: I1125 11:52:57.090376 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6kf6\" (UniqueName: \"kubernetes.io/projected/4b178e2c-b39e-4b45-afec-c81005eddebc-kube-api-access-r6kf6\") pod \"crc-debug-98lgp\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:57 crc kubenswrapper[4696]: I1125 11:52:57.174214 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:52:57 crc kubenswrapper[4696]: I1125 11:52:57.670410 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-98lgp" event={"ID":"4b178e2c-b39e-4b45-afec-c81005eddebc","Type":"ContainerStarted","Data":"7fc8de8928154931a2bb6bb25576d9cfd5d6e848cf86c86f4143e92f753c21b5"} Nov 25 11:53:00 crc kubenswrapper[4696]: I1125 11:53:00.801912 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:53:00 crc kubenswrapper[4696]: I1125 11:53:00.801973 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:53:00 crc kubenswrapper[4696]: I1125 11:53:00.802022 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 11:53:00 crc kubenswrapper[4696]: I1125 11:53:00.802828 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:53:00 crc kubenswrapper[4696]: I1125 11:53:00.802904 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" gracePeriod=600 Nov 25 11:53:00 crc kubenswrapper[4696]: E1125 11:53:00.956349 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:53:01 crc kubenswrapper[4696]: I1125 11:53:01.719150 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" exitCode=0 Nov 25 11:53:01 crc kubenswrapper[4696]: I1125 11:53:01.719221 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b"} Nov 25 11:53:01 crc kubenswrapper[4696]: I1125 11:53:01.719644 4696 scope.go:117] "RemoveContainer" containerID="0b65c8a742c0d66f761404eda52c203cf3b833d07f3569d491d2bdd59d6f3b40" Nov 25 11:53:01 crc kubenswrapper[4696]: I1125 11:53:01.720415 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:53:01 crc kubenswrapper[4696]: E1125 11:53:01.720697 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:53:09 crc kubenswrapper[4696]: I1125 11:53:09.811629 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-98lgp" event={"ID":"4b178e2c-b39e-4b45-afec-c81005eddebc","Type":"ContainerStarted","Data":"36ee53217bd86d448e0965f72d80e1fcf9621adfada82122da6a98c9b5613dc3"} Nov 25 11:53:09 crc kubenswrapper[4696]: I1125 11:53:09.827170 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-szjd4/crc-debug-98lgp" podStartSLOduration=2.183089368 podStartE2EDuration="13.827153359s" podCreationTimestamp="2025-11-25 11:52:56 +0000 UTC" firstStartedPulling="2025-11-25 11:52:57.207481492 +0000 UTC m=+4574.021098089" lastFinishedPulling="2025-11-25 11:53:08.851545493 +0000 UTC m=+4585.665162080" observedRunningTime="2025-11-25 11:53:09.823172742 +0000 UTC m=+4586.636789329" watchObservedRunningTime="2025-11-25 11:53:09.827153359 +0000 UTC m=+4586.640769946" Nov 25 11:53:14 crc kubenswrapper[4696]: I1125 11:53:14.050891 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:53:14 crc kubenswrapper[4696]: E1125 11:53:14.051888 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:53:25 crc kubenswrapper[4696]: I1125 11:53:25.883801 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-llfmt"] Nov 25 11:53:25 crc kubenswrapper[4696]: I1125 11:53:25.888116 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:25 crc kubenswrapper[4696]: I1125 11:53:25.894243 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-llfmt"] Nov 25 11:53:25 crc kubenswrapper[4696]: I1125 11:53:25.988257 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29brq\" (UniqueName: \"kubernetes.io/projected/ff7ff079-c37c-4973-852c-3c6afbadcda5-kube-api-access-29brq\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:25 crc kubenswrapper[4696]: I1125 11:53:25.988324 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7ff079-c37c-4973-852c-3c6afbadcda5-catalog-content\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:25 crc kubenswrapper[4696]: I1125 11:53:25.988904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7ff079-c37c-4973-852c-3c6afbadcda5-utilities\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.093042 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7ff079-c37c-4973-852c-3c6afbadcda5-catalog-content\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.093183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7ff079-c37c-4973-852c-3c6afbadcda5-utilities\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.093262 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29brq\" (UniqueName: \"kubernetes.io/projected/ff7ff079-c37c-4973-852c-3c6afbadcda5-kube-api-access-29brq\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.094160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7ff079-c37c-4973-852c-3c6afbadcda5-catalog-content\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.094975 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7ff079-c37c-4973-852c-3c6afbadcda5-utilities\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.116815 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29brq\" (UniqueName: \"kubernetes.io/projected/ff7ff079-c37c-4973-852c-3c6afbadcda5-kube-api-access-29brq\") pod \"redhat-operators-llfmt\" (UID: \"ff7ff079-c37c-4973-852c-3c6afbadcda5\") " pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.219638 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:26 crc kubenswrapper[4696]: I1125 11:53:26.847541 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-llfmt"] Nov 25 11:53:28 crc kubenswrapper[4696]: I1125 11:53:27.999683 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff7ff079-c37c-4973-852c-3c6afbadcda5" containerID="6818b4acff8a8a66f4c8948a1cb602c6bc287256be956d40246ef2600c51676d" exitCode=0 Nov 25 11:53:28 crc kubenswrapper[4696]: I1125 11:53:28.000283 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-llfmt" event={"ID":"ff7ff079-c37c-4973-852c-3c6afbadcda5","Type":"ContainerDied","Data":"6818b4acff8a8a66f4c8948a1cb602c6bc287256be956d40246ef2600c51676d"} Nov 25 11:53:28 crc kubenswrapper[4696]: I1125 11:53:28.000316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-llfmt" event={"ID":"ff7ff079-c37c-4973-852c-3c6afbadcda5","Type":"ContainerStarted","Data":"85f6b01d879febea23f5c4f3caede7277e52868069b0f69bd8c3d8ec63aa9a09"} Nov 25 11:53:28 crc kubenswrapper[4696]: I1125 11:53:28.042399 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:53:28 crc kubenswrapper[4696]: E1125 11:53:28.042725 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.254905 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-msmh5"] Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.260748 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.301164 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msmh5"] Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.313545 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-utilities\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.313699 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6lgg\" (UniqueName: \"kubernetes.io/projected/97007fcc-c15d-4d0e-ba71-e911135bd173-kube-api-access-r6lgg\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.313932 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-catalog-content\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.416061 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-utilities\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.416162 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6lgg\" (UniqueName: \"kubernetes.io/projected/97007fcc-c15d-4d0e-ba71-e911135bd173-kube-api-access-r6lgg\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.416302 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-catalog-content\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.417037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-catalog-content\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.417065 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-utilities\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:30 crc kubenswrapper[4696]: I1125 11:53:30.437780 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6lgg\" (UniqueName: \"kubernetes.io/projected/97007fcc-c15d-4d0e-ba71-e911135bd173-kube-api-access-r6lgg\") pod \"redhat-marketplace-msmh5\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:33 crc kubenswrapper[4696]: I1125 11:53:33.252828 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:33 crc kubenswrapper[4696]: I1125 11:53:33.760217 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msmh5"] Nov 25 11:53:39 crc kubenswrapper[4696]: I1125 11:53:39.042817 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:53:39 crc kubenswrapper[4696]: E1125 11:53:39.043605 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:53:39 crc kubenswrapper[4696]: I1125 11:53:39.107369 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerStarted","Data":"63608d8097adee33a6f42a10f7187160c829aa3dec60a36ebe6e38e25f19ecab"} Nov 25 11:53:40 crc kubenswrapper[4696]: I1125 11:53:40.120871 4696 generic.go:334] "Generic (PLEG): container finished" podID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerID="b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e" exitCode=0 Nov 25 11:53:40 crc kubenswrapper[4696]: I1125 11:53:40.121025 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerDied","Data":"b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e"} Nov 25 11:53:40 crc kubenswrapper[4696]: I1125 11:53:40.129253 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-llfmt" event={"ID":"ff7ff079-c37c-4973-852c-3c6afbadcda5","Type":"ContainerStarted","Data":"f84285941515cf833ae9c3c3587d5594f6feb0475091db65ff55fab387a447ae"} Nov 25 11:53:41 crc kubenswrapper[4696]: I1125 11:53:41.144920 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff7ff079-c37c-4973-852c-3c6afbadcda5" containerID="f84285941515cf833ae9c3c3587d5594f6feb0475091db65ff55fab387a447ae" exitCode=0 Nov 25 11:53:41 crc kubenswrapper[4696]: I1125 11:53:41.145233 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-llfmt" event={"ID":"ff7ff079-c37c-4973-852c-3c6afbadcda5","Type":"ContainerDied","Data":"f84285941515cf833ae9c3c3587d5594f6feb0475091db65ff55fab387a447ae"} Nov 25 11:53:44 crc kubenswrapper[4696]: I1125 11:53:44.188378 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerStarted","Data":"3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0"} Nov 25 11:53:44 crc kubenswrapper[4696]: I1125 11:53:44.190753 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-llfmt" event={"ID":"ff7ff079-c37c-4973-852c-3c6afbadcda5","Type":"ContainerStarted","Data":"be17e6ff16e30cccf0a2cb3f7c6b3a56f32667ad6cfe8a9e4c59584d29585726"} Nov 25 11:53:44 crc kubenswrapper[4696]: I1125 11:53:44.238258 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-llfmt" podStartSLOduration=4.12038551 podStartE2EDuration="19.238234281s" podCreationTimestamp="2025-11-25 11:53:25 +0000 UTC" firstStartedPulling="2025-11-25 11:53:28.091969753 +0000 UTC m=+4604.905586340" lastFinishedPulling="2025-11-25 11:53:43.209818534 +0000 UTC m=+4620.023435111" observedRunningTime="2025-11-25 11:53:44.227253863 +0000 UTC m=+4621.040870450" watchObservedRunningTime="2025-11-25 11:53:44.238234281 +0000 UTC m=+4621.051850868" Nov 25 11:53:45 crc kubenswrapper[4696]: I1125 11:53:45.202690 4696 generic.go:334] "Generic (PLEG): container finished" podID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerID="3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0" exitCode=0 Nov 25 11:53:45 crc kubenswrapper[4696]: I1125 11:53:45.202902 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerDied","Data":"3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0"} Nov 25 11:53:46 crc kubenswrapper[4696]: I1125 11:53:46.219971 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:46 crc kubenswrapper[4696]: I1125 11:53:46.220319 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:53:46 crc kubenswrapper[4696]: I1125 11:53:46.224570 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerStarted","Data":"df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf"} Nov 25 11:53:46 crc kubenswrapper[4696]: I1125 11:53:46.255502 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-msmh5" podStartSLOduration=10.738521181 podStartE2EDuration="16.255478702s" podCreationTimestamp="2025-11-25 11:53:30 +0000 UTC" firstStartedPulling="2025-11-25 11:53:40.123548801 +0000 UTC m=+4616.937165388" lastFinishedPulling="2025-11-25 11:53:45.640506322 +0000 UTC m=+4622.454122909" observedRunningTime="2025-11-25 11:53:46.246613721 +0000 UTC m=+4623.060230308" watchObservedRunningTime="2025-11-25 11:53:46.255478702 +0000 UTC m=+4623.069095289" Nov 25 11:53:47 crc kubenswrapper[4696]: I1125 11:53:47.299024 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-llfmt" podUID="ff7ff079-c37c-4973-852c-3c6afbadcda5" containerName="registry-server" probeResult="failure" output=< Nov 25 11:53:47 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:53:47 crc kubenswrapper[4696]: > Nov 25 11:53:51 crc kubenswrapper[4696]: I1125 11:53:51.042441 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:53:51 crc kubenswrapper[4696]: E1125 11:53:51.043392 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:53:53 crc kubenswrapper[4696]: I1125 11:53:53.252995 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:53 crc kubenswrapper[4696]: I1125 11:53:53.253276 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:53 crc kubenswrapper[4696]: I1125 11:53:53.323921 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:53 crc kubenswrapper[4696]: I1125 11:53:53.375527 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:53 crc kubenswrapper[4696]: I1125 11:53:53.567429 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msmh5"] Nov 25 11:53:55 crc kubenswrapper[4696]: I1125 11:53:55.315009 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-msmh5" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="registry-server" containerID="cri-o://df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf" gracePeriod=2 Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.101799 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.155828 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6lgg\" (UniqueName: \"kubernetes.io/projected/97007fcc-c15d-4d0e-ba71-e911135bd173-kube-api-access-r6lgg\") pod \"97007fcc-c15d-4d0e-ba71-e911135bd173\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.155986 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-catalog-content\") pod \"97007fcc-c15d-4d0e-ba71-e911135bd173\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.156024 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-utilities\") pod \"97007fcc-c15d-4d0e-ba71-e911135bd173\" (UID: \"97007fcc-c15d-4d0e-ba71-e911135bd173\") " Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.158152 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-utilities" (OuterVolumeSpecName: "utilities") pod "97007fcc-c15d-4d0e-ba71-e911135bd173" (UID: "97007fcc-c15d-4d0e-ba71-e911135bd173"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.159560 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.165860 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97007fcc-c15d-4d0e-ba71-e911135bd173-kube-api-access-r6lgg" (OuterVolumeSpecName: "kube-api-access-r6lgg") pod "97007fcc-c15d-4d0e-ba71-e911135bd173" (UID: "97007fcc-c15d-4d0e-ba71-e911135bd173"). InnerVolumeSpecName "kube-api-access-r6lgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.182619 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97007fcc-c15d-4d0e-ba71-e911135bd173" (UID: "97007fcc-c15d-4d0e-ba71-e911135bd173"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.261208 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6lgg\" (UniqueName: \"kubernetes.io/projected/97007fcc-c15d-4d0e-ba71-e911135bd173-kube-api-access-r6lgg\") on node \"crc\" DevicePath \"\"" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.261242 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97007fcc-c15d-4d0e-ba71-e911135bd173-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.328422 4696 generic.go:334] "Generic (PLEG): container finished" podID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerID="df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf" exitCode=0 Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.328461 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerDied","Data":"df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf"} Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.328487 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msmh5" event={"ID":"97007fcc-c15d-4d0e-ba71-e911135bd173","Type":"ContainerDied","Data":"63608d8097adee33a6f42a10f7187160c829aa3dec60a36ebe6e38e25f19ecab"} Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.328505 4696 scope.go:117] "RemoveContainer" containerID="df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.328649 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msmh5" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.358325 4696 scope.go:117] "RemoveContainer" containerID="3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.378005 4696 scope.go:117] "RemoveContainer" containerID="b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.384591 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msmh5"] Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.396981 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-msmh5"] Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.481174 4696 scope.go:117] "RemoveContainer" containerID="df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf" Nov 25 11:53:56 crc kubenswrapper[4696]: E1125 11:53:56.484118 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf\": container with ID starting with df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf not found: ID does not exist" containerID="df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.484156 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf"} err="failed to get container status \"df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf\": rpc error: code = NotFound desc = could not find container \"df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf\": container with ID starting with df4704b2d457b55df3126d99662f6fed7a9eec9ab52f5614730a7b7ab5a5dcdf not found: ID does not exist" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.484180 4696 scope.go:117] "RemoveContainer" containerID="3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0" Nov 25 11:53:56 crc kubenswrapper[4696]: E1125 11:53:56.484576 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0\": container with ID starting with 3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0 not found: ID does not exist" containerID="3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.484605 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0"} err="failed to get container status \"3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0\": rpc error: code = NotFound desc = could not find container \"3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0\": container with ID starting with 3b24fb5e4ada1f9706b8ae93a2956f5bc25f5b88566377ace82ee0daff88efa0 not found: ID does not exist" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.484624 4696 scope.go:117] "RemoveContainer" containerID="b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e" Nov 25 11:53:56 crc kubenswrapper[4696]: E1125 11:53:56.485041 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e\": container with ID starting with b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e not found: ID does not exist" containerID="b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e" Nov 25 11:53:56 crc kubenswrapper[4696]: I1125 11:53:56.485075 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e"} err="failed to get container status \"b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e\": rpc error: code = NotFound desc = could not find container \"b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e\": container with ID starting with b5a2081b8dc6e19e68706a2b6dbef5ef1de6d80c14ceda8a0e1e378e25f82b0e not found: ID does not exist" Nov 25 11:53:57 crc kubenswrapper[4696]: I1125 11:53:57.268571 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-llfmt" podUID="ff7ff079-c37c-4973-852c-3c6afbadcda5" containerName="registry-server" probeResult="failure" output=< Nov 25 11:53:57 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 11:53:57 crc kubenswrapper[4696]: > Nov 25 11:53:58 crc kubenswrapper[4696]: I1125 11:53:58.071285 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" path="/var/lib/kubelet/pods/97007fcc-c15d-4d0e-ba71-e911135bd173/volumes" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.894503 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fdzjm"] Nov 25 11:54:05 crc kubenswrapper[4696]: E1125 11:54:05.895522 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="extract-content" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.895541 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="extract-content" Nov 25 11:54:05 crc kubenswrapper[4696]: E1125 11:54:05.895571 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="registry-server" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.895579 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="registry-server" Nov 25 11:54:05 crc kubenswrapper[4696]: E1125 11:54:05.895599 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="extract-utilities" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.895608 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="extract-utilities" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.899218 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="97007fcc-c15d-4d0e-ba71-e911135bd173" containerName="registry-server" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.901234 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.906846 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fdzjm"] Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.912486 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-utilities\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.912618 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-catalog-content\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:05 crc kubenswrapper[4696]: I1125 11:54:05.912723 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clcxx\" (UniqueName: \"kubernetes.io/projected/88a0f0f8-846e-4f27-aa6a-699edde896d6-kube-api-access-clcxx\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.030406 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clcxx\" (UniqueName: \"kubernetes.io/projected/88a0f0f8-846e-4f27-aa6a-699edde896d6-kube-api-access-clcxx\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.030572 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-utilities\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.030651 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-catalog-content\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.031168 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-catalog-content\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.031252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-utilities\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.043309 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:54:06 crc kubenswrapper[4696]: E1125 11:54:06.043538 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.053895 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clcxx\" (UniqueName: \"kubernetes.io/projected/88a0f0f8-846e-4f27-aa6a-699edde896d6-kube-api-access-clcxx\") pod \"certified-operators-fdzjm\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.096289 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f8lhf"] Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.098702 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.125220 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8lhf"] Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.147327 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-utilities\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.147554 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-catalog-content\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.147826 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvbkw\" (UniqueName: \"kubernetes.io/projected/9998684b-0335-4777-8483-9fbf14acf9f8-kube-api-access-rvbkw\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.236981 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.249463 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-utilities\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.249579 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-catalog-content\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.249663 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvbkw\" (UniqueName: \"kubernetes.io/projected/9998684b-0335-4777-8483-9fbf14acf9f8-kube-api-access-rvbkw\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.250085 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-utilities\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.250317 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-catalog-content\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.279408 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvbkw\" (UniqueName: \"kubernetes.io/projected/9998684b-0335-4777-8483-9fbf14acf9f8-kube-api-access-rvbkw\") pod \"community-operators-f8lhf\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.321750 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.395394 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-llfmt" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.465615 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:06 crc kubenswrapper[4696]: I1125 11:54:06.749557 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fdzjm"] Nov 25 11:54:07 crc kubenswrapper[4696]: I1125 11:54:07.279677 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8lhf"] Nov 25 11:54:07 crc kubenswrapper[4696]: W1125 11:54:07.305683 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9998684b_0335_4777_8483_9fbf14acf9f8.slice/crio-a14b1a476b8b011656f3b19dd8dc97199f799481d13e5edea43fa39c10d79280 WatchSource:0}: Error finding container a14b1a476b8b011656f3b19dd8dc97199f799481d13e5edea43fa39c10d79280: Status 404 returned error can't find the container with id a14b1a476b8b011656f3b19dd8dc97199f799481d13e5edea43fa39c10d79280 Nov 25 11:54:07 crc kubenswrapper[4696]: I1125 11:54:07.456241 4696 generic.go:334] "Generic (PLEG): container finished" podID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerID="93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be" exitCode=0 Nov 25 11:54:07 crc kubenswrapper[4696]: I1125 11:54:07.456305 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerDied","Data":"93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be"} Nov 25 11:54:07 crc kubenswrapper[4696]: I1125 11:54:07.456331 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerStarted","Data":"574c60a29b43619c643036ded7108e4f2279ec3612a9dd7bd26c69f898d6861b"} Nov 25 11:54:07 crc kubenswrapper[4696]: I1125 11:54:07.457584 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerStarted","Data":"a14b1a476b8b011656f3b19dd8dc97199f799481d13e5edea43fa39c10d79280"} Nov 25 11:54:08 crc kubenswrapper[4696]: I1125 11:54:08.467130 4696 generic.go:334] "Generic (PLEG): container finished" podID="9998684b-0335-4777-8483-9fbf14acf9f8" containerID="a6ec6bbe22b7d3bd32175ed73434ccc0a279bef928f201474d6ecdf388702a1e" exitCode=0 Nov 25 11:54:08 crc kubenswrapper[4696]: I1125 11:54:08.467286 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerDied","Data":"a6ec6bbe22b7d3bd32175ed73434ccc0a279bef928f201474d6ecdf388702a1e"} Nov 25 11:54:09 crc kubenswrapper[4696]: I1125 11:54:09.489295 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerStarted","Data":"cfe2671f84e41ed45fd6b0e3b68559864f3c5f3b7fbd7793c5170a517a8e334d"} Nov 25 11:54:10 crc kubenswrapper[4696]: I1125 11:54:10.108707 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-llfmt"] Nov 25 11:54:10 crc kubenswrapper[4696]: I1125 11:54:10.489748 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:54:10 crc kubenswrapper[4696]: I1125 11:54:10.490276 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2jlsf" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" containerID="cri-o://6a3fbd239bc0a33da320e2096db29761194952aa2e0e3bac52bdae2c33a22b63" gracePeriod=2 Nov 25 11:54:11 crc kubenswrapper[4696]: I1125 11:54:11.523441 4696 generic.go:334] "Generic (PLEG): container finished" podID="9998684b-0335-4777-8483-9fbf14acf9f8" containerID="cfe2671f84e41ed45fd6b0e3b68559864f3c5f3b7fbd7793c5170a517a8e334d" exitCode=0 Nov 25 11:54:11 crc kubenswrapper[4696]: I1125 11:54:11.523619 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerDied","Data":"cfe2671f84e41ed45fd6b0e3b68559864f3c5f3b7fbd7793c5170a517a8e334d"} Nov 25 11:54:11 crc kubenswrapper[4696]: I1125 11:54:11.529483 4696 generic.go:334] "Generic (PLEG): container finished" podID="a4329258-c066-423c-b084-db2267dddec8" containerID="6a3fbd239bc0a33da320e2096db29761194952aa2e0e3bac52bdae2c33a22b63" exitCode=0 Nov 25 11:54:11 crc kubenswrapper[4696]: I1125 11:54:11.529538 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerDied","Data":"6a3fbd239bc0a33da320e2096db29761194952aa2e0e3bac52bdae2c33a22b63"} Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.066800 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.133391 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-utilities\") pod \"a4329258-c066-423c-b084-db2267dddec8\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.133586 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-catalog-content\") pod \"a4329258-c066-423c-b084-db2267dddec8\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.133963 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkm2w\" (UniqueName: \"kubernetes.io/projected/a4329258-c066-423c-b084-db2267dddec8-kube-api-access-mkm2w\") pod \"a4329258-c066-423c-b084-db2267dddec8\" (UID: \"a4329258-c066-423c-b084-db2267dddec8\") " Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.136107 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-utilities" (OuterVolumeSpecName: "utilities") pod "a4329258-c066-423c-b084-db2267dddec8" (UID: "a4329258-c066-423c-b084-db2267dddec8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.148839 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4329258-c066-423c-b084-db2267dddec8-kube-api-access-mkm2w" (OuterVolumeSpecName: "kube-api-access-mkm2w") pod "a4329258-c066-423c-b084-db2267dddec8" (UID: "a4329258-c066-423c-b084-db2267dddec8"). InnerVolumeSpecName "kube-api-access-mkm2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.237860 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkm2w\" (UniqueName: \"kubernetes.io/projected/a4329258-c066-423c-b084-db2267dddec8-kube-api-access-mkm2w\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.238110 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.311092 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4329258-c066-423c-b084-db2267dddec8" (UID: "a4329258-c066-423c-b084-db2267dddec8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.340238 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4329258-c066-423c-b084-db2267dddec8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.540069 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerStarted","Data":"2ba619c864c03a483e99401b0e6f0bb0c9161bbbe0a67c3d79e27c68860e30ca"} Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.545163 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlsf" event={"ID":"a4329258-c066-423c-b084-db2267dddec8","Type":"ContainerDied","Data":"e3bcb41e1eeb20dbd59e754eed34a15d72207410d4b69f8dea04a884229534ea"} Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.545369 4696 scope.go:117] "RemoveContainer" containerID="6a3fbd239bc0a33da320e2096db29761194952aa2e0e3bac52bdae2c33a22b63" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.545558 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlsf" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.561271 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f8lhf" podStartSLOduration=3.097606842 podStartE2EDuration="6.561250162s" podCreationTimestamp="2025-11-25 11:54:06 +0000 UTC" firstStartedPulling="2025-11-25 11:54:08.469295952 +0000 UTC m=+4645.282912539" lastFinishedPulling="2025-11-25 11:54:11.932939282 +0000 UTC m=+4648.746555859" observedRunningTime="2025-11-25 11:54:12.560099256 +0000 UTC m=+4649.373715863" watchObservedRunningTime="2025-11-25 11:54:12.561250162 +0000 UTC m=+4649.374866749" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.566504 4696 scope.go:117] "RemoveContainer" containerID="4b535fe280cb6c88274216f56652de2aae7cdbc3f1533816c022550fb1f9ebe5" Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.597401 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.608414 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2jlsf"] Nov 25 11:54:12 crc kubenswrapper[4696]: I1125 11:54:12.844802 4696 scope.go:117] "RemoveContainer" containerID="8b4926f1a525b52937e1c90ddead1b04a851f6ae6ab96a641245873ce715baaa" Nov 25 11:54:14 crc kubenswrapper[4696]: I1125 11:54:14.064336 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4329258-c066-423c-b084-db2267dddec8" path="/var/lib/kubelet/pods/a4329258-c066-423c-b084-db2267dddec8/volumes" Nov 25 11:54:15 crc kubenswrapper[4696]: I1125 11:54:15.578564 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerStarted","Data":"9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400"} Nov 25 11:54:16 crc kubenswrapper[4696]: I1125 11:54:16.466834 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:16 crc kubenswrapper[4696]: I1125 11:54:16.467153 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:16 crc kubenswrapper[4696]: I1125 11:54:16.544909 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:17 crc kubenswrapper[4696]: I1125 11:54:17.595206 4696 generic.go:334] "Generic (PLEG): container finished" podID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerID="9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400" exitCode=0 Nov 25 11:54:17 crc kubenswrapper[4696]: I1125 11:54:17.595273 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerDied","Data":"9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400"} Nov 25 11:54:18 crc kubenswrapper[4696]: I1125 11:54:18.607914 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerStarted","Data":"39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e"} Nov 25 11:54:18 crc kubenswrapper[4696]: I1125 11:54:18.629147 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fdzjm" podStartSLOduration=3.101135288 podStartE2EDuration="13.629129184s" podCreationTimestamp="2025-11-25 11:54:05 +0000 UTC" firstStartedPulling="2025-11-25 11:54:07.458468732 +0000 UTC m=+4644.272085319" lastFinishedPulling="2025-11-25 11:54:17.986462628 +0000 UTC m=+4654.800079215" observedRunningTime="2025-11-25 11:54:18.627239923 +0000 UTC m=+4655.440856510" watchObservedRunningTime="2025-11-25 11:54:18.629129184 +0000 UTC m=+4655.442745771" Nov 25 11:54:19 crc kubenswrapper[4696]: I1125 11:54:19.043128 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:54:19 crc kubenswrapper[4696]: E1125 11:54:19.043446 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:54:25 crc kubenswrapper[4696]: I1125 11:54:25.689341 4696 generic.go:334] "Generic (PLEG): container finished" podID="4b178e2c-b39e-4b45-afec-c81005eddebc" containerID="36ee53217bd86d448e0965f72d80e1fcf9621adfada82122da6a98c9b5613dc3" exitCode=0 Nov 25 11:54:25 crc kubenswrapper[4696]: I1125 11:54:25.689627 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-98lgp" event={"ID":"4b178e2c-b39e-4b45-afec-c81005eddebc","Type":"ContainerDied","Data":"36ee53217bd86d448e0965f72d80e1fcf9621adfada82122da6a98c9b5613dc3"} Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.238072 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.238869 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.291270 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.520043 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.751878 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.825189 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.866174 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-szjd4/crc-debug-98lgp"] Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.874849 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-szjd4/crc-debug-98lgp"] Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.940922 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6kf6\" (UniqueName: \"kubernetes.io/projected/4b178e2c-b39e-4b45-afec-c81005eddebc-kube-api-access-r6kf6\") pod \"4b178e2c-b39e-4b45-afec-c81005eddebc\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.941089 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b178e2c-b39e-4b45-afec-c81005eddebc-host\") pod \"4b178e2c-b39e-4b45-afec-c81005eddebc\" (UID: \"4b178e2c-b39e-4b45-afec-c81005eddebc\") " Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.941219 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b178e2c-b39e-4b45-afec-c81005eddebc-host" (OuterVolumeSpecName: "host") pod "4b178e2c-b39e-4b45-afec-c81005eddebc" (UID: "4b178e2c-b39e-4b45-afec-c81005eddebc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.942035 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b178e2c-b39e-4b45-afec-c81005eddebc-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:26 crc kubenswrapper[4696]: I1125 11:54:26.949291 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b178e2c-b39e-4b45-afec-c81005eddebc-kube-api-access-r6kf6" (OuterVolumeSpecName: "kube-api-access-r6kf6") pod "4b178e2c-b39e-4b45-afec-c81005eddebc" (UID: "4b178e2c-b39e-4b45-afec-c81005eddebc"). InnerVolumeSpecName "kube-api-access-r6kf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.043502 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6kf6\" (UniqueName: \"kubernetes.io/projected/4b178e2c-b39e-4b45-afec-c81005eddebc-kube-api-access-r6kf6\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.303652 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8lhf"] Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.303991 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f8lhf" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="registry-server" containerID="cri-o://2ba619c864c03a483e99401b0e6f0bb0c9161bbbe0a67c3d79e27c68860e30ca" gracePeriod=2 Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.713001 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fc8de8928154931a2bb6bb25576d9cfd5d6e848cf86c86f4143e92f753c21b5" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.713016 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-98lgp" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.716535 4696 generic.go:334] "Generic (PLEG): container finished" podID="9998684b-0335-4777-8483-9fbf14acf9f8" containerID="2ba619c864c03a483e99401b0e6f0bb0c9161bbbe0a67c3d79e27c68860e30ca" exitCode=0 Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.717588 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerDied","Data":"2ba619c864c03a483e99401b0e6f0bb0c9161bbbe0a67c3d79e27c68860e30ca"} Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.717619 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8lhf" event={"ID":"9998684b-0335-4777-8483-9fbf14acf9f8","Type":"ContainerDied","Data":"a14b1a476b8b011656f3b19dd8dc97199f799481d13e5edea43fa39c10d79280"} Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.717701 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a14b1a476b8b011656f3b19dd8dc97199f799481d13e5edea43fa39c10d79280" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.814114 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.871540 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-catalog-content\") pod \"9998684b-0335-4777-8483-9fbf14acf9f8\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.871789 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-utilities\") pod \"9998684b-0335-4777-8483-9fbf14acf9f8\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.871820 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvbkw\" (UniqueName: \"kubernetes.io/projected/9998684b-0335-4777-8483-9fbf14acf9f8-kube-api-access-rvbkw\") pod \"9998684b-0335-4777-8483-9fbf14acf9f8\" (UID: \"9998684b-0335-4777-8483-9fbf14acf9f8\") " Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.874764 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-utilities" (OuterVolumeSpecName: "utilities") pod "9998684b-0335-4777-8483-9fbf14acf9f8" (UID: "9998684b-0335-4777-8483-9fbf14acf9f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.878046 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9998684b-0335-4777-8483-9fbf14acf9f8-kube-api-access-rvbkw" (OuterVolumeSpecName: "kube-api-access-rvbkw") pod "9998684b-0335-4777-8483-9fbf14acf9f8" (UID: "9998684b-0335-4777-8483-9fbf14acf9f8"). InnerVolumeSpecName "kube-api-access-rvbkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.927838 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9998684b-0335-4777-8483-9fbf14acf9f8" (UID: "9998684b-0335-4777-8483-9fbf14acf9f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.975025 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.975066 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9998684b-0335-4777-8483-9fbf14acf9f8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:27 crc kubenswrapper[4696]: I1125 11:54:27.975077 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvbkw\" (UniqueName: \"kubernetes.io/projected/9998684b-0335-4777-8483-9fbf14acf9f8-kube-api-access-rvbkw\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.060987 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b178e2c-b39e-4b45-afec-c81005eddebc" path="/var/lib/kubelet/pods/4b178e2c-b39e-4b45-afec-c81005eddebc/volumes" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133206 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-szjd4/crc-debug-txfzp"] Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133751 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="extract-utilities" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133773 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="extract-utilities" Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133797 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133806 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133827 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="extract-content" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133836 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="extract-content" Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133863 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="extract-content" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133874 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="extract-content" Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133890 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="registry-server" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133899 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="registry-server" Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133916 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="extract-utilities" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133938 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="extract-utilities" Nov 25 11:54:28 crc kubenswrapper[4696]: E1125 11:54:28.133953 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b178e2c-b39e-4b45-afec-c81005eddebc" containerName="container-00" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.133961 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b178e2c-b39e-4b45-afec-c81005eddebc" containerName="container-00" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.134205 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" containerName="registry-server" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.134233 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4329258-c066-423c-b084-db2267dddec8" containerName="registry-server" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.134247 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b178e2c-b39e-4b45-afec-c81005eddebc" containerName="container-00" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.135108 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.180165 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78bbec2e-a6f8-4847-8119-9560fe62a5de-host\") pod \"crc-debug-txfzp\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.180413 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8vhv\" (UniqueName: \"kubernetes.io/projected/78bbec2e-a6f8-4847-8119-9560fe62a5de-kube-api-access-h8vhv\") pod \"crc-debug-txfzp\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.282518 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78bbec2e-a6f8-4847-8119-9560fe62a5de-host\") pod \"crc-debug-txfzp\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.282657 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78bbec2e-a6f8-4847-8119-9560fe62a5de-host\") pod \"crc-debug-txfzp\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.282809 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8vhv\" (UniqueName: \"kubernetes.io/projected/78bbec2e-a6f8-4847-8119-9560fe62a5de-kube-api-access-h8vhv\") pod \"crc-debug-txfzp\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.300536 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8vhv\" (UniqueName: \"kubernetes.io/projected/78bbec2e-a6f8-4847-8119-9560fe62a5de-kube-api-access-h8vhv\") pod \"crc-debug-txfzp\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.457581 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.726304 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-txfzp" event={"ID":"78bbec2e-a6f8-4847-8119-9560fe62a5de","Type":"ContainerStarted","Data":"1ff2982196ab48c6d88d9e20c3deeac311bd2b393a6f40d008d3501d635d57f5"} Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.726353 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-txfzp" event={"ID":"78bbec2e-a6f8-4847-8119-9560fe62a5de","Type":"ContainerStarted","Data":"b33106ba71bbaf4473098e211928b5dabef44e32a5df6a6223027de89d167779"} Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.726407 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8lhf" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.757628 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-szjd4/crc-debug-txfzp" podStartSLOduration=0.757609054 podStartE2EDuration="757.609054ms" podCreationTimestamp="2025-11-25 11:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:54:28.752292186 +0000 UTC m=+4665.565908773" watchObservedRunningTime="2025-11-25 11:54:28.757609054 +0000 UTC m=+4665.571225641" Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.786277 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8lhf"] Nov 25 11:54:28 crc kubenswrapper[4696]: I1125 11:54:28.796540 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f8lhf"] Nov 25 11:54:29 crc kubenswrapper[4696]: I1125 11:54:29.102122 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fdzjm"] Nov 25 11:54:29 crc kubenswrapper[4696]: I1125 11:54:29.739503 4696 generic.go:334] "Generic (PLEG): container finished" podID="78bbec2e-a6f8-4847-8119-9560fe62a5de" containerID="1ff2982196ab48c6d88d9e20c3deeac311bd2b393a6f40d008d3501d635d57f5" exitCode=0 Nov 25 11:54:29 crc kubenswrapper[4696]: I1125 11:54:29.739606 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-txfzp" event={"ID":"78bbec2e-a6f8-4847-8119-9560fe62a5de","Type":"ContainerDied","Data":"1ff2982196ab48c6d88d9e20c3deeac311bd2b393a6f40d008d3501d635d57f5"} Nov 25 11:54:29 crc kubenswrapper[4696]: I1125 11:54:29.740423 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fdzjm" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="registry-server" containerID="cri-o://39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e" gracePeriod=2 Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.042836 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:54:30 crc kubenswrapper[4696]: E1125 11:54:30.043682 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.066490 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9998684b-0335-4777-8483-9fbf14acf9f8" path="/var/lib/kubelet/pods/9998684b-0335-4777-8483-9fbf14acf9f8/volumes" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.258595 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.330609 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-catalog-content\") pod \"88a0f0f8-846e-4f27-aa6a-699edde896d6\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.331107 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clcxx\" (UniqueName: \"kubernetes.io/projected/88a0f0f8-846e-4f27-aa6a-699edde896d6-kube-api-access-clcxx\") pod \"88a0f0f8-846e-4f27-aa6a-699edde896d6\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.331168 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-utilities\") pod \"88a0f0f8-846e-4f27-aa6a-699edde896d6\" (UID: \"88a0f0f8-846e-4f27-aa6a-699edde896d6\") " Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.332912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-utilities" (OuterVolumeSpecName: "utilities") pod "88a0f0f8-846e-4f27-aa6a-699edde896d6" (UID: "88a0f0f8-846e-4f27-aa6a-699edde896d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.339089 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a0f0f8-846e-4f27-aa6a-699edde896d6-kube-api-access-clcxx" (OuterVolumeSpecName: "kube-api-access-clcxx") pod "88a0f0f8-846e-4f27-aa6a-699edde896d6" (UID: "88a0f0f8-846e-4f27-aa6a-699edde896d6"). InnerVolumeSpecName "kube-api-access-clcxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.412052 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88a0f0f8-846e-4f27-aa6a-699edde896d6" (UID: "88a0f0f8-846e-4f27-aa6a-699edde896d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.434063 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.434102 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a0f0f8-846e-4f27-aa6a-699edde896d6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.434116 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clcxx\" (UniqueName: \"kubernetes.io/projected/88a0f0f8-846e-4f27-aa6a-699edde896d6-kube-api-access-clcxx\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.748772 4696 generic.go:334] "Generic (PLEG): container finished" podID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerID="39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e" exitCode=0 Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.748848 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdzjm" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.748971 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerDied","Data":"39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e"} Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.749016 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdzjm" event={"ID":"88a0f0f8-846e-4f27-aa6a-699edde896d6","Type":"ContainerDied","Data":"574c60a29b43619c643036ded7108e4f2279ec3612a9dd7bd26c69f898d6861b"} Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.749037 4696 scope.go:117] "RemoveContainer" containerID="39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.837959 4696 scope.go:117] "RemoveContainer" containerID="9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.838645 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.854983 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fdzjm"] Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.866083 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fdzjm"] Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.907117 4696 scope.go:117] "RemoveContainer" containerID="93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.941081 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8vhv\" (UniqueName: \"kubernetes.io/projected/78bbec2e-a6f8-4847-8119-9560fe62a5de-kube-api-access-h8vhv\") pod \"78bbec2e-a6f8-4847-8119-9560fe62a5de\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.941134 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78bbec2e-a6f8-4847-8119-9560fe62a5de-host\") pod \"78bbec2e-a6f8-4847-8119-9560fe62a5de\" (UID: \"78bbec2e-a6f8-4847-8119-9560fe62a5de\") " Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.941516 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78bbec2e-a6f8-4847-8119-9560fe62a5de-host" (OuterVolumeSpecName: "host") pod "78bbec2e-a6f8-4847-8119-9560fe62a5de" (UID: "78bbec2e-a6f8-4847-8119-9560fe62a5de"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.941763 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78bbec2e-a6f8-4847-8119-9560fe62a5de-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.945681 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78bbec2e-a6f8-4847-8119-9560fe62a5de-kube-api-access-h8vhv" (OuterVolumeSpecName: "kube-api-access-h8vhv") pod "78bbec2e-a6f8-4847-8119-9560fe62a5de" (UID: "78bbec2e-a6f8-4847-8119-9560fe62a5de"). InnerVolumeSpecName "kube-api-access-h8vhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.974298 4696 scope.go:117] "RemoveContainer" containerID="39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e" Nov 25 11:54:30 crc kubenswrapper[4696]: E1125 11:54:30.974691 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e\": container with ID starting with 39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e not found: ID does not exist" containerID="39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.974729 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e"} err="failed to get container status \"39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e\": rpc error: code = NotFound desc = could not find container \"39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e\": container with ID starting with 39e54b8c7351468d72be9982cac0ec30c381d79f17d4811fee12a1bd9604075e not found: ID does not exist" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.974753 4696 scope.go:117] "RemoveContainer" containerID="9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400" Nov 25 11:54:30 crc kubenswrapper[4696]: E1125 11:54:30.975021 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400\": container with ID starting with 9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400 not found: ID does not exist" containerID="9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.975051 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400"} err="failed to get container status \"9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400\": rpc error: code = NotFound desc = could not find container \"9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400\": container with ID starting with 9425cf897424b2dfdf37fd1b628a244a298166b619664f999559bc7b67463400 not found: ID does not exist" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.975070 4696 scope.go:117] "RemoveContainer" containerID="93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be" Nov 25 11:54:30 crc kubenswrapper[4696]: E1125 11:54:30.975478 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be\": container with ID starting with 93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be not found: ID does not exist" containerID="93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be" Nov 25 11:54:30 crc kubenswrapper[4696]: I1125 11:54:30.975508 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be"} err="failed to get container status \"93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be\": rpc error: code = NotFound desc = could not find container \"93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be\": container with ID starting with 93af2c986c93e40c150bcf7eda1aa35ccfac2c40291158c45766816f6a3900be not found: ID does not exist" Nov 25 11:54:31 crc kubenswrapper[4696]: I1125 11:54:31.043278 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8vhv\" (UniqueName: \"kubernetes.io/projected/78bbec2e-a6f8-4847-8119-9560fe62a5de-kube-api-access-h8vhv\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:31 crc kubenswrapper[4696]: I1125 11:54:31.427256 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-szjd4/crc-debug-txfzp"] Nov 25 11:54:31 crc kubenswrapper[4696]: I1125 11:54:31.436147 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-szjd4/crc-debug-txfzp"] Nov 25 11:54:31 crc kubenswrapper[4696]: I1125 11:54:31.760433 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b33106ba71bbaf4473098e211928b5dabef44e32a5df6a6223027de89d167779" Nov 25 11:54:31 crc kubenswrapper[4696]: I1125 11:54:31.760525 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-txfzp" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.054075 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78bbec2e-a6f8-4847-8119-9560fe62a5de" path="/var/lib/kubelet/pods/78bbec2e-a6f8-4847-8119-9560fe62a5de/volumes" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.054747 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" path="/var/lib/kubelet/pods/88a0f0f8-846e-4f27-aa6a-699edde896d6/volumes" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.593845 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-szjd4/crc-debug-7wp7t"] Nov 25 11:54:32 crc kubenswrapper[4696]: E1125 11:54:32.594222 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="registry-server" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.594234 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="registry-server" Nov 25 11:54:32 crc kubenswrapper[4696]: E1125 11:54:32.594260 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78bbec2e-a6f8-4847-8119-9560fe62a5de" containerName="container-00" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.594265 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="78bbec2e-a6f8-4847-8119-9560fe62a5de" containerName="container-00" Nov 25 11:54:32 crc kubenswrapper[4696]: E1125 11:54:32.594290 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="extract-utilities" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.594297 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="extract-utilities" Nov 25 11:54:32 crc kubenswrapper[4696]: E1125 11:54:32.594307 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="extract-content" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.594312 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="extract-content" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.594482 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="78bbec2e-a6f8-4847-8119-9560fe62a5de" containerName="container-00" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.594496 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a0f0f8-846e-4f27-aa6a-699edde896d6" containerName="registry-server" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.595197 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.670587 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-host\") pod \"crc-debug-7wp7t\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.670934 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsj2b\" (UniqueName: \"kubernetes.io/projected/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-kube-api-access-tsj2b\") pod \"crc-debug-7wp7t\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.773714 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-host\") pod \"crc-debug-7wp7t\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.774257 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-host\") pod \"crc-debug-7wp7t\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.774282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsj2b\" (UniqueName: \"kubernetes.io/projected/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-kube-api-access-tsj2b\") pod \"crc-debug-7wp7t\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.800514 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsj2b\" (UniqueName: \"kubernetes.io/projected/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-kube-api-access-tsj2b\") pod \"crc-debug-7wp7t\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: I1125 11:54:32.910027 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:32 crc kubenswrapper[4696]: W1125 11:54:32.934223 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4cb66b_3a3f_4aad_b663_3f484cc1d9a5.slice/crio-1d0b84d0d04a757d2bdb119fdf6b3f3b30a4f5af42767c510239305665d06b7e WatchSource:0}: Error finding container 1d0b84d0d04a757d2bdb119fdf6b3f3b30a4f5af42767c510239305665d06b7e: Status 404 returned error can't find the container with id 1d0b84d0d04a757d2bdb119fdf6b3f3b30a4f5af42767c510239305665d06b7e Nov 25 11:54:33 crc kubenswrapper[4696]: I1125 11:54:33.779644 4696 generic.go:334] "Generic (PLEG): container finished" podID="2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" containerID="6f08a9e9a451c135c84ea16801744ca9b8003058c2b86426303c11e1fe642860" exitCode=0 Nov 25 11:54:33 crc kubenswrapper[4696]: I1125 11:54:33.779706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-7wp7t" event={"ID":"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5","Type":"ContainerDied","Data":"6f08a9e9a451c135c84ea16801744ca9b8003058c2b86426303c11e1fe642860"} Nov 25 11:54:33 crc kubenswrapper[4696]: I1125 11:54:33.780004 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/crc-debug-7wp7t" event={"ID":"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5","Type":"ContainerStarted","Data":"1d0b84d0d04a757d2bdb119fdf6b3f3b30a4f5af42767c510239305665d06b7e"} Nov 25 11:54:33 crc kubenswrapper[4696]: I1125 11:54:33.830509 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-szjd4/crc-debug-7wp7t"] Nov 25 11:54:33 crc kubenswrapper[4696]: I1125 11:54:33.845821 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-szjd4/crc-debug-7wp7t"] Nov 25 11:54:34 crc kubenswrapper[4696]: I1125 11:54:34.920164 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.018410 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-host\") pod \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.018508 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsj2b\" (UniqueName: \"kubernetes.io/projected/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-kube-api-access-tsj2b\") pod \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\" (UID: \"2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5\") " Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.019771 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-host" (OuterVolumeSpecName: "host") pod "2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" (UID: "2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.020461 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.037448 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-kube-api-access-tsj2b" (OuterVolumeSpecName: "kube-api-access-tsj2b") pod "2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" (UID: "2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5"). InnerVolumeSpecName "kube-api-access-tsj2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.122591 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsj2b\" (UniqueName: \"kubernetes.io/projected/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5-kube-api-access-tsj2b\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.814940 4696 scope.go:117] "RemoveContainer" containerID="6f08a9e9a451c135c84ea16801744ca9b8003058c2b86426303c11e1fe642860" Nov 25 11:54:35 crc kubenswrapper[4696]: I1125 11:54:35.814989 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/crc-debug-7wp7t" Nov 25 11:54:36 crc kubenswrapper[4696]: I1125 11:54:36.053409 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" path="/var/lib/kubelet/pods/2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5/volumes" Nov 25 11:54:42 crc kubenswrapper[4696]: I1125 11:54:42.042890 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:54:42 crc kubenswrapper[4696]: E1125 11:54:42.052318 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:54:52 crc kubenswrapper[4696]: I1125 11:54:52.610981 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9c4d748fd-g4t2c_c0f652ee-e64b-423b-902a-bab2bd12d46f/barbican-api/0.log" Nov 25 11:54:52 crc kubenswrapper[4696]: I1125 11:54:52.764281 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9c4d748fd-g4t2c_c0f652ee-e64b-423b-902a-bab2bd12d46f/barbican-api-log/0.log" Nov 25 11:54:52 crc kubenswrapper[4696]: I1125 11:54:52.877851 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-66cc7b58db-xph79_9e0f5310-cf60-4125-9b91-2dfda5383c7d/barbican-keystone-listener/0.log" Nov 25 11:54:52 crc kubenswrapper[4696]: I1125 11:54:52.977792 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-66cc7b58db-xph79_9e0f5310-cf60-4125-9b91-2dfda5383c7d/barbican-keystone-listener-log/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.120878 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64b77754f5-27gpf_22dbae09-113b-4136-b9d6-6162815b5ef7/barbican-worker/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.125284 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64b77754f5-27gpf_22dbae09-113b-4136-b9d6-6162815b5ef7/barbican-worker-log/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.259321 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-65952_062b8e49-6137-4b89-8ae4-a29272ff0d1f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.435321 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/ceilometer-central-agent/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.486541 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/ceilometer-notification-agent/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.517191 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/proxy-httpd/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.643549 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/sg-core/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.764721 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5a8e77cd-7f3d-45e6-bb36-2ce208f70adb/cinder-api/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.790590 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5a8e77cd-7f3d-45e6-bb36-2ce208f70adb/cinder-api-log/0.log" Nov 25 11:54:53 crc kubenswrapper[4696]: I1125 11:54:53.986585 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_dccb9177-738d-433c-a2ec-99131e690f89/cinder-scheduler/0.log" Nov 25 11:54:54 crc kubenswrapper[4696]: I1125 11:54:54.046866 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_dccb9177-738d-433c-a2ec-99131e690f89/probe/0.log" Nov 25 11:54:54 crc kubenswrapper[4696]: I1125 11:54:54.244160 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp_574f645b-d7d9-4756-82ef-27bfa7bf982e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:54 crc kubenswrapper[4696]: I1125 11:54:54.291900 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bzgww_114399b2-d253-44d9-994e-5098ef72a972/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:54 crc kubenswrapper[4696]: I1125 11:54:54.461538 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-qdbhs_6b74d08c-9100-49d2-8483-34d4c76c0c6b/init/0.log" Nov 25 11:54:54 crc kubenswrapper[4696]: I1125 11:54:54.675619 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-qdbhs_6b74d08c-9100-49d2-8483-34d4c76c0c6b/init/0.log" Nov 25 11:54:55 crc kubenswrapper[4696]: I1125 11:54:55.693379 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-j29t5_38409cac-958b-497e-9431-4d0cafe2fd8d/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:55 crc kubenswrapper[4696]: I1125 11:54:55.742177 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-qdbhs_6b74d08c-9100-49d2-8483-34d4c76c0c6b/dnsmasq-dns/0.log" Nov 25 11:54:55 crc kubenswrapper[4696]: I1125 11:54:55.895334 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d642739c-282a-4644-9aae-68f82c3ddf58/glance-httpd/0.log" Nov 25 11:54:55 crc kubenswrapper[4696]: I1125 11:54:55.957556 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d642739c-282a-4644-9aae-68f82c3ddf58/glance-log/0.log" Nov 25 11:54:56 crc kubenswrapper[4696]: I1125 11:54:56.015947 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_bd3e6e15-4a89-468d-8151-26c9e4c19dc5/glance-httpd/0.log" Nov 25 11:54:56 crc kubenswrapper[4696]: I1125 11:54:56.114434 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_bd3e6e15-4a89-468d-8151-26c9e4c19dc5/glance-log/0.log" Nov 25 11:54:56 crc kubenswrapper[4696]: I1125 11:54:56.253728 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6558fc797b-2qp4r_1944527b-6ffc-4b6b-954f-4b01394ea0cf/horizon/1.log" Nov 25 11:54:56 crc kubenswrapper[4696]: I1125 11:54:56.453477 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6558fc797b-2qp4r_1944527b-6ffc-4b6b-954f-4b01394ea0cf/horizon/0.log" Nov 25 11:54:56 crc kubenswrapper[4696]: I1125 11:54:56.682586 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6_3202ae74-b34c-41aa-b717-6313d9e71bf4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:56 crc kubenswrapper[4696]: I1125 11:54:56.894380 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6558fc797b-2qp4r_1944527b-6ffc-4b6b-954f-4b01394ea0cf/horizon-log/0.log" Nov 25 11:54:57 crc kubenswrapper[4696]: I1125 11:54:57.043021 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:54:57 crc kubenswrapper[4696]: E1125 11:54:57.043529 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:54:57 crc kubenswrapper[4696]: I1125 11:54:57.050968 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kg6v5_d69df5c8-e28a-4b73-9d2b-97087aaed351/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:57 crc kubenswrapper[4696]: I1125 11:54:57.558062 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401141-q8gxc_0a3da9a0-6262-4ff8-b510-e476db9d9985/keystone-cron/0.log" Nov 25 11:54:57 crc kubenswrapper[4696]: I1125 11:54:57.607293 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9ed1a610-3fb4-4643-a481-724f5036b836/kube-state-metrics/0.log" Nov 25 11:54:57 crc kubenswrapper[4696]: I1125 11:54:57.702242 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5547bcb8fc-b65gm_ad61fdbd-b908-4430-8b43-e08260f0dbd1/keystone-api/0.log" Nov 25 11:54:57 crc kubenswrapper[4696]: I1125 11:54:57.936755 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr_27d6547e-1c82-4b39-a281-476463f32e81/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:58 crc kubenswrapper[4696]: I1125 11:54:58.409770 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns_9813095c-910d-451b-b3bb-18a6d5fcdf26/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:54:58 crc kubenswrapper[4696]: I1125 11:54:58.674846 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c599bcb65-hgc2v_412cdec0-a9b1-4796-82f6-95c1f6b9b6ae/neutron-httpd/0.log" Nov 25 11:54:58 crc kubenswrapper[4696]: I1125 11:54:58.903863 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c599bcb65-hgc2v_412cdec0-a9b1-4796-82f6-95c1f6b9b6ae/neutron-api/0.log" Nov 25 11:54:59 crc kubenswrapper[4696]: I1125 11:54:59.473485 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8189ed6c-e99d-44db-bd26-7b74ea6539fb/nova-cell0-conductor-conductor/0.log" Nov 25 11:54:59 crc kubenswrapper[4696]: I1125 11:54:59.667946 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_984434ed-24d1-4090-a7d9-c38b39617b74/nova-cell1-conductor-conductor/0.log" Nov 25 11:54:59 crc kubenswrapper[4696]: I1125 11:54:59.782855 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_eba40268-f540-4ac7-9ae4-14455c81395f/memcached/0.log" Nov 25 11:55:00 crc kubenswrapper[4696]: I1125 11:55:00.131129 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_672502a7-5ef3-4596-9507-21afa27d44e6/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 11:55:00 crc kubenswrapper[4696]: I1125 11:55:00.217514 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8377dca6-a47c-4ffb-9916-f6505eed1e48/nova-api-log/0.log" Nov 25 11:55:00 crc kubenswrapper[4696]: I1125 11:55:00.235531 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-4fkn4_11902ad3-7187-4749-8fbf-3d5f1eb77764/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:00 crc kubenswrapper[4696]: I1125 11:55:00.391502 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8377dca6-a47c-4ffb-9916-f6505eed1e48/nova-api-api/0.log" Nov 25 11:55:00 crc kubenswrapper[4696]: I1125 11:55:00.458367 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d808ab02-7a75-4441-a5d0-b92fcaa1eb06/nova-metadata-log/0.log" Nov 25 11:55:00 crc kubenswrapper[4696]: I1125 11:55:00.832382 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c541fdf-4691-4f61-9865-8497d636667d/mysql-bootstrap/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.070109 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_97905b5e-c4ba-4712-8ed6-e1f1c228f683/nova-scheduler-scheduler/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.155153 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c541fdf-4691-4f61-9865-8497d636667d/galera/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.160913 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c541fdf-4691-4f61-9865-8497d636667d/mysql-bootstrap/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.379394 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ad9b8866-501c-48e8-a694-5fff74e631fc/mysql-bootstrap/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.685394 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_14829ad2-5154-4bc3-ac0d-8343c93c0f6d/openstackclient/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.758109 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ad9b8866-501c-48e8-a694-5fff74e631fc/galera/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.766216 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ad9b8866-501c-48e8-a694-5fff74e631fc/mysql-bootstrap/0.log" Nov 25 11:55:01 crc kubenswrapper[4696]: I1125 11:55:01.897910 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d808ab02-7a75-4441-a5d0-b92fcaa1eb06/nova-metadata-metadata/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.000873 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jz4h7_87dc2400-74ac-4b0b-ae7d-ca62010572d3/ovn-controller/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.044265 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-qld29_d900f0ea-8197-4383-bdbb-666b60db9f19/openstack-network-exporter/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.213422 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovsdb-server-init/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.338114 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovsdb-server-init/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.350344 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovsdb-server/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.422567 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovs-vswitchd/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.515094 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-zf8h7_348a4b45-dfe7-449f-8b03-871394bfba46/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.571281 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e0d794d3-7682-4248-affe-b53f4f34f264/openstack-network-exporter/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.689642 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e0d794d3-7682-4248-affe-b53f4f34f264/ovn-northd/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.716510 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8b4e55fc-7bf4-4a72-b805-b7b2776de33c/openstack-network-exporter/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.869831 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8b4e55fc-7bf4-4a72-b805-b7b2776de33c/ovsdbserver-nb/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.939327 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbc1467c-cf49-44a9-99eb-c592c38412a1/openstack-network-exporter/0.log" Nov 25 11:55:02 crc kubenswrapper[4696]: I1125 11:55:02.970512 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbc1467c-cf49-44a9-99eb-c592c38412a1/ovsdbserver-sb/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.210790 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf5b7831-caa9-4411-9386-cf99ba7f8a20/setup-container/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.230779 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-646fccfb54-pbhr5_537410db-3589-4c1a-bb28-1f0a66a20ff6/placement-api/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.430779 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-646fccfb54-pbhr5_537410db-3589-4c1a-bb28-1f0a66a20ff6/placement-log/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.446023 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf5b7831-caa9-4411-9386-cf99ba7f8a20/setup-container/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.511824 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf5b7831-caa9-4411-9386-cf99ba7f8a20/rabbitmq/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.606605 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b161351-5bdf-4cfe-895b-29381797d908/setup-container/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.819617 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b161351-5bdf-4cfe-895b-29381797d908/rabbitmq/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.833171 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7_5779201b-7609-43c1-9cec-c3cc15302542/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:03 crc kubenswrapper[4696]: I1125 11:55:03.875076 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b161351-5bdf-4cfe-895b-29381797d908/setup-container/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.031953 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ls6nj_2a97a4bb-2255-4f00-901a-3392d7808c53/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.152602 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx_4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.262535 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-w8rsq_438e2932-b4ff-4135-b4a8-7b4e89d574fc/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.328120 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9w2md_d97e4985-25b9-4106-8053-b7ee1aa42542/ssh-known-hosts-edpm-deployment/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.561962 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78446968f5-84c89_142ffa20-cd2a-4b6b-90ff-2770e6eed020/proxy-httpd/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.569654 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78446968f5-84c89_142ffa20-cd2a-4b6b-90ff-2770e6eed020/proxy-server/0.log" Nov 25 11:55:04 crc kubenswrapper[4696]: I1125 11:55:04.595450 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-xshf5_dbeda82d-3956-47f7-95a8-6ffc25961e5d/swift-ring-rebalance/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.188323 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-auditor/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.370459 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-reaper/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.381083 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-server/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.418710 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-auditor/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.438941 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-replicator/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.497159 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-replicator/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.575072 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-updater/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.613521 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-server/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.704049 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-auditor/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.721400 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-expirer/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.729007 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-replicator/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.824619 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-server/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.888031 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-updater/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.913113 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/swift-recon-cron/0.log" Nov 25 11:55:05 crc kubenswrapper[4696]: I1125 11:55:05.975644 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/rsync/0.log" Nov 25 11:55:06 crc kubenswrapper[4696]: I1125 11:55:06.105858 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz_372198a7-8311-4133-8047-fcb9908b7f90/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:06 crc kubenswrapper[4696]: I1125 11:55:06.184281 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_f8d03f21-fad7-432e-8220-12094b55a1a1/tempest-tests-tempest-tests-runner/0.log" Nov 25 11:55:06 crc kubenswrapper[4696]: I1125 11:55:06.285474 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6/test-operator-logs-container/0.log" Nov 25 11:55:06 crc kubenswrapper[4696]: I1125 11:55:06.901406 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s_b756844d-cdfd-4345-8a32-035b71161bfc/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:55:11 crc kubenswrapper[4696]: I1125 11:55:11.042102 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:55:11 crc kubenswrapper[4696]: E1125 11:55:11.043805 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:55:26 crc kubenswrapper[4696]: I1125 11:55:26.043285 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:55:26 crc kubenswrapper[4696]: E1125 11:55:26.046506 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.204425 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/util/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.389068 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/pull/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.398593 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/util/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.401896 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/pull/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.680971 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/pull/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.692420 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/extract/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.757863 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/util/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.924902 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-4kpwn_c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd/kube-rbac-proxy/0.log" Nov 25 11:55:33 crc kubenswrapper[4696]: I1125 11:55:33.938210 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-4kpwn_c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd/manager/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.015939 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-5kdhx_900bb9ad-0f16-4eb4-beb5-c23e0089db39/kube-rbac-proxy/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.223871 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-5kdhx_900bb9ad-0f16-4eb4-beb5-c23e0089db39/manager/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.235168 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-vrf9f_2ae50512-8525-4878-80b1-7e65a4c035d9/kube-rbac-proxy/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.331281 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-vrf9f_2ae50512-8525-4878-80b1-7e65a4c035d9/manager/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.472706 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-ldvdx_7844a3a7-9ef3-48b7-9ab8-944304690567/kube-rbac-proxy/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.563526 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-ldvdx_7844a3a7-9ef3-48b7-9ab8-944304690567/manager/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.747583 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-c7l6c_28a8de6d-a7e1-4ca9-853d-6d27a066df12/kube-rbac-proxy/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.801295 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-c7l6c_28a8de6d-a7e1-4ca9-853d-6d27a066df12/manager/0.log" Nov 25 11:55:34 crc kubenswrapper[4696]: I1125 11:55:34.912512 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2p89s_347bfeb5-b62b-4515-be16-1fdb0fc52e44/kube-rbac-proxy/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.033449 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2p89s_347bfeb5-b62b-4515-be16-1fdb0fc52e44/manager/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.088950 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-2xtrg_f8dc76df-0e30-435e-b61e-926c2b60a87e/kube-rbac-proxy/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.364627 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q4ml6_a891b8d2-92ed-4f76-9947-3968ee474921/manager/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.384821 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-2xtrg_f8dc76df-0e30-435e-b61e-926c2b60a87e/manager/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.416147 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q4ml6_a891b8d2-92ed-4f76-9947-3968ee474921/kube-rbac-proxy/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.712759 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-4fkts_01eb2add-76e2-414f-8a70-275d2a1b8939/manager/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.738804 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-4fkts_01eb2add-76e2-414f-8a70-275d2a1b8939/kube-rbac-proxy/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.845361 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-xwzk7_ce501661-3cfd-4cce-8256-c347638da2f6/kube-rbac-proxy/0.log" Nov 25 11:55:35 crc kubenswrapper[4696]: I1125 11:55:35.918718 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-xwzk7_ce501661-3cfd-4cce-8256-c347638da2f6/manager/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.009118 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-vdjlx_394e9c9d-8d79-4be4-9679-432c658bd52d/kube-rbac-proxy/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.126278 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-vdjlx_394e9c9d-8d79-4be4-9679-432c658bd52d/manager/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.247556 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-xqqtv_8faec67a-3309-496b-9c5a-74eb6421a7f2/kube-rbac-proxy/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.334855 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-xqqtv_8faec67a-3309-496b-9c5a-74eb6421a7f2/manager/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.446121 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-7xgqt_c40524c0-d668-4b1c-8113-67b152b150e7/kube-rbac-proxy/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.693850 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-7xgqt_c40524c0-d668-4b1c-8113-67b152b150e7/manager/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.802954 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-gjl4t_c6ca26a1-b284-48b2-b68c-fd0c83b404ec/kube-rbac-proxy/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.804624 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-gjl4t_c6ca26a1-b284-48b2-b68c-fd0c83b404ec/manager/0.log" Nov 25 11:55:36 crc kubenswrapper[4696]: I1125 11:55:36.959872 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t_e6871bcf-31f0-4782-b94c-56e1002cd0d1/kube-rbac-proxy/0.log" Nov 25 11:55:37 crc kubenswrapper[4696]: I1125 11:55:37.029829 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t_e6871bcf-31f0-4782-b94c-56e1002cd0d1/manager/0.log" Nov 25 11:55:37 crc kubenswrapper[4696]: I1125 11:55:37.453772 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cccc9d798-j4nfn_62e71f26-481d-4d1f-b3c7-bed2f91968cd/operator/0.log" Nov 25 11:55:37 crc kubenswrapper[4696]: I1125 11:55:37.602774 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-55w9h_2d7b9c54-45bc-4bc8-9513-a37f199b022f/registry-server/0.log" Nov 25 11:55:37 crc kubenswrapper[4696]: I1125 11:55:37.988790 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-599kv_7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f/manager/0.log" Nov 25 11:55:37 crc kubenswrapper[4696]: I1125 11:55:37.994303 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-599kv_7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f/kube-rbac-proxy/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.217574 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-gnbb8_33329628-2334-4d45-ad9f-20d2d23ab8ba/kube-rbac-proxy/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.350180 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-gnbb8_33329628-2334-4d45-ad9f-20d2d23ab8ba/manager/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.411547 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94k5m_3367886a-6027-418f-9415-38d775304079/operator/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.657879 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65d98ff956-rrwtq_ad6baa72-3698-484a-a2f2-690719d71677/manager/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.695985 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-9ln5q_33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2/kube-rbac-proxy/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.831781 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-9ln5q_33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2/manager/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.913044 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-jfrk4_90aa0c65-64c1-4ec3-a104-122691247673/kube-rbac-proxy/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.927107 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-q2c8n_3f468970-3367-47b7-82fa-e9af018c3996/kube-rbac-proxy/0.log" Nov 25 11:55:38 crc kubenswrapper[4696]: I1125 11:55:38.973179 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-jfrk4_90aa0c65-64c1-4ec3-a104-122691247673/manager/0.log" Nov 25 11:55:39 crc kubenswrapper[4696]: I1125 11:55:39.148749 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-q2c8n_3f468970-3367-47b7-82fa-e9af018c3996/manager/0.log" Nov 25 11:55:39 crc kubenswrapper[4696]: I1125 11:55:39.197692 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-758cd_17643fa7-c097-4a6c-ac8a-fe7ec0cc9901/kube-rbac-proxy/0.log" Nov 25 11:55:39 crc kubenswrapper[4696]: I1125 11:55:39.327198 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-758cd_17643fa7-c097-4a6c-ac8a-fe7ec0cc9901/manager/0.log" Nov 25 11:55:40 crc kubenswrapper[4696]: I1125 11:55:40.042304 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:55:40 crc kubenswrapper[4696]: E1125 11:55:40.042770 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:55:51 crc kubenswrapper[4696]: I1125 11:55:51.043253 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:55:51 crc kubenswrapper[4696]: E1125 11:55:51.044087 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:55:59 crc kubenswrapper[4696]: I1125 11:55:59.928600 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-n62qw_2f6f8fdc-fec0-46b9-9277-2996da698f24/control-plane-machine-set-operator/0.log" Nov 25 11:56:00 crc kubenswrapper[4696]: I1125 11:56:00.132753 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5g9hp_23a01a08-d008-451f-86c0-f3c456fafb88/kube-rbac-proxy/0.log" Nov 25 11:56:00 crc kubenswrapper[4696]: I1125 11:56:00.174681 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5g9hp_23a01a08-d008-451f-86c0-f3c456fafb88/machine-api-operator/0.log" Nov 25 11:56:06 crc kubenswrapper[4696]: I1125 11:56:06.042302 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:56:06 crc kubenswrapper[4696]: E1125 11:56:06.043202 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:56:17 crc kubenswrapper[4696]: I1125 11:56:17.003377 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-dvr5m_c4daf7cb-c6fd-43da-8821-d26fe9f88fac/cert-manager-controller/0.log" Nov 25 11:56:17 crc kubenswrapper[4696]: I1125 11:56:17.042366 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:56:17 crc kubenswrapper[4696]: E1125 11:56:17.042844 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:56:17 crc kubenswrapper[4696]: I1125 11:56:17.084777 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qt985_c1f44937-600c-4df9-94f4-8adbc7a9b0ef/cert-manager-cainjector/0.log" Nov 25 11:56:17 crc kubenswrapper[4696]: I1125 11:56:17.292140 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-knqmh_ed104e46-ddf7-496c-a389-1c78d24e8e7c/cert-manager-webhook/0.log" Nov 25 11:56:29 crc kubenswrapper[4696]: I1125 11:56:29.042874 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:56:29 crc kubenswrapper[4696]: E1125 11:56:29.043656 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:56:30 crc kubenswrapper[4696]: I1125 11:56:30.912027 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-52cq5_b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332/nmstate-console-plugin/0.log" Nov 25 11:56:31 crc kubenswrapper[4696]: I1125 11:56:31.080551 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ttnxf_a2a61ba7-423c-4fbc-9bf5-63af41ee1170/nmstate-handler/0.log" Nov 25 11:56:31 crc kubenswrapper[4696]: I1125 11:56:31.203833 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-xznfr_585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d/nmstate-metrics/0.log" Nov 25 11:56:31 crc kubenswrapper[4696]: I1125 11:56:31.215696 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-xznfr_585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d/kube-rbac-proxy/0.log" Nov 25 11:56:31 crc kubenswrapper[4696]: I1125 11:56:31.432465 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-wkpl2_d246ee4c-dd53-42a6-b24f-5e34687b6c8e/nmstate-operator/0.log" Nov 25 11:56:31 crc kubenswrapper[4696]: I1125 11:56:31.507020 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-5bpk8_fc74e56a-3f49-4d38-b5e4-78cafb3276a6/nmstate-webhook/0.log" Nov 25 11:56:42 crc kubenswrapper[4696]: I1125 11:56:42.042290 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:56:42 crc kubenswrapper[4696]: E1125 11:56:42.043083 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.081376 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-jgmgn_b24625b7-df3d-4f50-ba97-5eac956b1662/kube-rbac-proxy/0.log" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.274369 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-jgmgn_b24625b7-df3d-4f50-ba97-5eac956b1662/controller/0.log" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.350630 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.649493 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.755398 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.767828 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 11:56:49 crc kubenswrapper[4696]: I1125 11:56:49.799610 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.408238 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.443836 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.467538 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.490023 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.696121 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.698173 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.717765 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/controller/0.log" Nov 25 11:56:50 crc kubenswrapper[4696]: I1125 11:56:50.767657 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 11:56:51 crc kubenswrapper[4696]: I1125 11:56:51.046801 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/frr-metrics/0.log" Nov 25 11:56:51 crc kubenswrapper[4696]: I1125 11:56:51.048581 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/kube-rbac-proxy/0.log" Nov 25 11:56:51 crc kubenswrapper[4696]: I1125 11:56:51.166158 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/kube-rbac-proxy-frr/0.log" Nov 25 11:56:51 crc kubenswrapper[4696]: I1125 11:56:51.395096 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/reloader/0.log" Nov 25 11:56:51 crc kubenswrapper[4696]: I1125 11:56:51.536321 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-nd8g6_c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f/frr-k8s-webhook-server/0.log" Nov 25 11:56:51 crc kubenswrapper[4696]: I1125 11:56:51.866203 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d5845fdbd-84v4v_5ad9cb9e-8e80-41bc-bb0a-687f580830b0/manager/0.log" Nov 25 11:56:52 crc kubenswrapper[4696]: I1125 11:56:52.150532 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b88d4b6cb-m8sz5_c8010a11-0748-4b8b-b40d-14d57fe53f40/webhook-server/0.log" Nov 25 11:56:52 crc kubenswrapper[4696]: I1125 11:56:52.224103 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/frr/0.log" Nov 25 11:56:52 crc kubenswrapper[4696]: I1125 11:56:52.363261 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42dc9_e4999207-9162-4635-b7a8-2bd0d5311019/kube-rbac-proxy/0.log" Nov 25 11:56:52 crc kubenswrapper[4696]: I1125 11:56:52.799472 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42dc9_e4999207-9162-4635-b7a8-2bd0d5311019/speaker/0.log" Nov 25 11:56:56 crc kubenswrapper[4696]: I1125 11:56:56.043138 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:56:56 crc kubenswrapper[4696]: E1125 11:56:56.043987 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:57:08 crc kubenswrapper[4696]: I1125 11:57:08.043215 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:57:08 crc kubenswrapper[4696]: E1125 11:57:08.043835 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:57:08 crc kubenswrapper[4696]: I1125 11:57:08.515131 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/util/0.log" Nov 25 11:57:08 crc kubenswrapper[4696]: I1125 11:57:08.726479 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/util/0.log" Nov 25 11:57:08 crc kubenswrapper[4696]: I1125 11:57:08.766412 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/pull/0.log" Nov 25 11:57:08 crc kubenswrapper[4696]: I1125 11:57:08.770761 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/pull/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.035007 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/util/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.076987 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/pull/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.151988 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/extract/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.243349 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-utilities/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.440017 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-utilities/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.471525 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-content/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.507151 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-content/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.717593 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-content/0.log" Nov 25 11:57:09 crc kubenswrapper[4696]: I1125 11:57:09.741709 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-utilities/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.008792 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-utilities/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.201878 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/registry-server/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.385289 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-utilities/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.432147 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-content/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.454901 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-content/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.751173 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-utilities/0.log" Nov 25 11:57:10 crc kubenswrapper[4696]: I1125 11:57:10.794791 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-content/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.125788 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/util/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.534051 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/util/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.612520 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/pull/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.626818 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/pull/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.901704 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/pull/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.931352 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/util/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.935075 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/registry-server/0.log" Nov 25 11:57:11 crc kubenswrapper[4696]: I1125 11:57:11.949435 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/extract/0.log" Nov 25 11:57:12 crc kubenswrapper[4696]: I1125 11:57:12.186310 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xrn8c_e77b3a6e-c1e9-44ef-806f-6adcba39fe31/marketplace-operator/1.log" Nov 25 11:57:12 crc kubenswrapper[4696]: I1125 11:57:12.243176 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xrn8c_e77b3a6e-c1e9-44ef-806f-6adcba39fe31/marketplace-operator/0.log" Nov 25 11:57:12 crc kubenswrapper[4696]: I1125 11:57:12.400526 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-utilities/0.log" Nov 25 11:57:12 crc kubenswrapper[4696]: I1125 11:57:12.617157 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-utilities/0.log" Nov 25 11:57:12 crc kubenswrapper[4696]: I1125 11:57:12.647384 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-content/0.log" Nov 25 11:57:12 crc kubenswrapper[4696]: I1125 11:57:12.692816 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-content/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.204808 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-content/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.276609 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-utilities/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.441555 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-utilities/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.442873 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/registry-server/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.524361 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-content/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.602860 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-utilities/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.641819 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-content/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.843482 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-utilities/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.847450 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-content/0.log" Nov 25 11:57:13 crc kubenswrapper[4696]: I1125 11:57:13.989271 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/registry-server/0.log" Nov 25 11:57:21 crc kubenswrapper[4696]: I1125 11:57:21.042831 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:57:21 crc kubenswrapper[4696]: E1125 11:57:21.043518 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:57:36 crc kubenswrapper[4696]: I1125 11:57:36.042877 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:57:36 crc kubenswrapper[4696]: E1125 11:57:36.043511 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:57:47 crc kubenswrapper[4696]: I1125 11:57:47.042070 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:57:47 crc kubenswrapper[4696]: E1125 11:57:47.042845 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:57:58 crc kubenswrapper[4696]: I1125 11:57:58.042853 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:57:58 crc kubenswrapper[4696]: E1125 11:57:58.043698 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 11:58:12 crc kubenswrapper[4696]: I1125 11:58:12.042476 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 11:58:13 crc kubenswrapper[4696]: I1125 11:58:13.028285 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"332ab86e2b889db1ce953dd2bf5d3716c0deb9acbbf4dfb647276bf2b6d5b2b2"} Nov 25 11:59:37 crc kubenswrapper[4696]: I1125 11:59:37.931302 4696 scope.go:117] "RemoveContainer" containerID="36ee53217bd86d448e0965f72d80e1fcf9621adfada82122da6a98c9b5613dc3" Nov 25 11:59:47 crc kubenswrapper[4696]: I1125 11:59:47.936442 4696 generic.go:334] "Generic (PLEG): container finished" podID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerID="2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0" exitCode=0 Nov 25 11:59:47 crc kubenswrapper[4696]: I1125 11:59:47.936578 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-szjd4/must-gather-lg95d" event={"ID":"60f42251-7fe6-4fdb-aaec-fbafe9c9647c","Type":"ContainerDied","Data":"2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0"} Nov 25 11:59:47 crc kubenswrapper[4696]: I1125 11:59:47.937556 4696 scope.go:117] "RemoveContainer" containerID="2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0" Nov 25 11:59:48 crc kubenswrapper[4696]: I1125 11:59:48.987523 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-szjd4_must-gather-lg95d_60f42251-7fe6-4fdb-aaec-fbafe9c9647c/gather/0.log" Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.109975 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-szjd4/must-gather-lg95d"] Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.110790 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-szjd4/must-gather-lg95d" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="copy" containerID="cri-o://f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4" gracePeriod=2 Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.119953 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-szjd4/must-gather-lg95d"] Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.647499 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-szjd4_must-gather-lg95d_60f42251-7fe6-4fdb-aaec-fbafe9c9647c/copy/0.log" Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.648803 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.781769 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvp7v\" (UniqueName: \"kubernetes.io/projected/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-kube-api-access-hvp7v\") pod \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.781897 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-must-gather-output\") pod \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\" (UID: \"60f42251-7fe6-4fdb-aaec-fbafe9c9647c\") " Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.849102 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-kube-api-access-hvp7v" (OuterVolumeSpecName: "kube-api-access-hvp7v") pod "60f42251-7fe6-4fdb-aaec-fbafe9c9647c" (UID: "60f42251-7fe6-4fdb-aaec-fbafe9c9647c"). InnerVolumeSpecName "kube-api-access-hvp7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.886597 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvp7v\" (UniqueName: \"kubernetes.io/projected/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-kube-api-access-hvp7v\") on node \"crc\" DevicePath \"\"" Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.953832 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "60f42251-7fe6-4fdb-aaec-fbafe9c9647c" (UID: "60f42251-7fe6-4fdb-aaec-fbafe9c9647c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:59:57 crc kubenswrapper[4696]: I1125 11:59:57.988114 4696 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60f42251-7fe6-4fdb-aaec-fbafe9c9647c-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.055442 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" path="/var/lib/kubelet/pods/60f42251-7fe6-4fdb-aaec-fbafe9c9647c/volumes" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.056069 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-szjd4_must-gather-lg95d_60f42251-7fe6-4fdb-aaec-fbafe9c9647c/copy/0.log" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.056568 4696 generic.go:334] "Generic (PLEG): container finished" podID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerID="f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4" exitCode=143 Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.056702 4696 scope.go:117] "RemoveContainer" containerID="f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.056708 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-szjd4/must-gather-lg95d" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.082272 4696 scope.go:117] "RemoveContainer" containerID="2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.158481 4696 scope.go:117] "RemoveContainer" containerID="f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4" Nov 25 11:59:58 crc kubenswrapper[4696]: E1125 11:59:58.160157 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4\": container with ID starting with f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4 not found: ID does not exist" containerID="f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.160189 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4"} err="failed to get container status \"f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4\": rpc error: code = NotFound desc = could not find container \"f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4\": container with ID starting with f11e6be042b5be6e1ffff816e1465a230ae22499e69f40bbab5000f7713bb9b4 not found: ID does not exist" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.160210 4696 scope.go:117] "RemoveContainer" containerID="2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0" Nov 25 11:59:58 crc kubenswrapper[4696]: E1125 11:59:58.160784 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0\": container with ID starting with 2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0 not found: ID does not exist" containerID="2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0" Nov 25 11:59:58 crc kubenswrapper[4696]: I1125 11:59:58.160839 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0"} err="failed to get container status \"2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0\": rpc error: code = NotFound desc = could not find container \"2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0\": container with ID starting with 2ef8c3399a0d4bb131b68f769e5b681e45935c11464cdaa283e35feba3ba33a0 not found: ID does not exist" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.179038 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2"] Nov 25 12:00:00 crc kubenswrapper[4696]: E1125 12:00:00.179788 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="gather" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.179806 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="gather" Nov 25 12:00:00 crc kubenswrapper[4696]: E1125 12:00:00.179824 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="copy" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.179833 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="copy" Nov 25 12:00:00 crc kubenswrapper[4696]: E1125 12:00:00.179844 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" containerName="container-00" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.179852 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" containerName="container-00" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.180093 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4cb66b-3a3f-4aad-b663-3f484cc1d9a5" containerName="container-00" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.180125 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="copy" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.180137 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f42251-7fe6-4fdb-aaec-fbafe9c9647c" containerName="gather" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.181945 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.188314 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.188707 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.198653 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2"] Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.239337 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kcqg\" (UniqueName: \"kubernetes.io/projected/680e9982-47c3-43b3-ab5e-4d83de22aaea-kube-api-access-2kcqg\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.239417 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/680e9982-47c3-43b3-ab5e-4d83de22aaea-secret-volume\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.239471 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/680e9982-47c3-43b3-ab5e-4d83de22aaea-config-volume\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.341393 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kcqg\" (UniqueName: \"kubernetes.io/projected/680e9982-47c3-43b3-ab5e-4d83de22aaea-kube-api-access-2kcqg\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.342144 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/680e9982-47c3-43b3-ab5e-4d83de22aaea-secret-volume\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.343175 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/680e9982-47c3-43b3-ab5e-4d83de22aaea-config-volume\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.344059 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/680e9982-47c3-43b3-ab5e-4d83de22aaea-config-volume\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.350298 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/680e9982-47c3-43b3-ab5e-4d83de22aaea-secret-volume\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.363272 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kcqg\" (UniqueName: \"kubernetes.io/projected/680e9982-47c3-43b3-ab5e-4d83de22aaea-kube-api-access-2kcqg\") pod \"collect-profiles-29401200-j42t2\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:00 crc kubenswrapper[4696]: I1125 12:00:00.509464 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:01 crc kubenswrapper[4696]: I1125 12:00:01.053186 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2"] Nov 25 12:00:01 crc kubenswrapper[4696]: I1125 12:00:01.104628 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" event={"ID":"680e9982-47c3-43b3-ab5e-4d83de22aaea","Type":"ContainerStarted","Data":"ceaf9634f37279375b54521bb487c27f8f2a6538643148b02188962dea0dbda0"} Nov 25 12:00:02 crc kubenswrapper[4696]: I1125 12:00:02.125109 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" event={"ID":"680e9982-47c3-43b3-ab5e-4d83de22aaea","Type":"ContainerStarted","Data":"88128fe8c3b70c8cc45d9c0980c503accee1c788a2a61196fb82a5ffcd2d87e4"} Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.157614 4696 generic.go:334] "Generic (PLEG): container finished" podID="680e9982-47c3-43b3-ab5e-4d83de22aaea" containerID="88128fe8c3b70c8cc45d9c0980c503accee1c788a2a61196fb82a5ffcd2d87e4" exitCode=0 Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.157936 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" event={"ID":"680e9982-47c3-43b3-ab5e-4d83de22aaea","Type":"ContainerDied","Data":"88128fe8c3b70c8cc45d9c0980c503accee1c788a2a61196fb82a5ffcd2d87e4"} Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.540176 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.620636 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/680e9982-47c3-43b3-ab5e-4d83de22aaea-config-volume\") pod \"680e9982-47c3-43b3-ab5e-4d83de22aaea\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.620776 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/680e9982-47c3-43b3-ab5e-4d83de22aaea-secret-volume\") pod \"680e9982-47c3-43b3-ab5e-4d83de22aaea\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.620907 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kcqg\" (UniqueName: \"kubernetes.io/projected/680e9982-47c3-43b3-ab5e-4d83de22aaea-kube-api-access-2kcqg\") pod \"680e9982-47c3-43b3-ab5e-4d83de22aaea\" (UID: \"680e9982-47c3-43b3-ab5e-4d83de22aaea\") " Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.622468 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680e9982-47c3-43b3-ab5e-4d83de22aaea-config-volume" (OuterVolumeSpecName: "config-volume") pod "680e9982-47c3-43b3-ab5e-4d83de22aaea" (UID: "680e9982-47c3-43b3-ab5e-4d83de22aaea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.632258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/680e9982-47c3-43b3-ab5e-4d83de22aaea-kube-api-access-2kcqg" (OuterVolumeSpecName: "kube-api-access-2kcqg") pod "680e9982-47c3-43b3-ab5e-4d83de22aaea" (UID: "680e9982-47c3-43b3-ab5e-4d83de22aaea"). InnerVolumeSpecName "kube-api-access-2kcqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.632274 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/680e9982-47c3-43b3-ab5e-4d83de22aaea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "680e9982-47c3-43b3-ab5e-4d83de22aaea" (UID: "680e9982-47c3-43b3-ab5e-4d83de22aaea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.723220 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/680e9982-47c3-43b3-ab5e-4d83de22aaea-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.723458 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/680e9982-47c3-43b3-ab5e-4d83de22aaea-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:03 crc kubenswrapper[4696]: I1125 12:00:03.723527 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kcqg\" (UniqueName: \"kubernetes.io/projected/680e9982-47c3-43b3-ab5e-4d83de22aaea-kube-api-access-2kcqg\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:04 crc kubenswrapper[4696]: I1125 12:00:04.169541 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" event={"ID":"680e9982-47c3-43b3-ab5e-4d83de22aaea","Type":"ContainerDied","Data":"ceaf9634f37279375b54521bb487c27f8f2a6538643148b02188962dea0dbda0"} Nov 25 12:00:04 crc kubenswrapper[4696]: I1125 12:00:04.169906 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceaf9634f37279375b54521bb487c27f8f2a6538643148b02188962dea0dbda0" Nov 25 12:00:04 crc kubenswrapper[4696]: I1125 12:00:04.169849 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-j42t2" Nov 25 12:00:04 crc kubenswrapper[4696]: I1125 12:00:04.644855 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd"] Nov 25 12:00:04 crc kubenswrapper[4696]: I1125 12:00:04.654060 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-z22qd"] Nov 25 12:00:06 crc kubenswrapper[4696]: I1125 12:00:06.065525 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a64ab406-d59c-4369-9ca9-91c47ee6887e" path="/var/lib/kubelet/pods/a64ab406-d59c-4369-9ca9-91c47ee6887e/volumes" Nov 25 12:00:30 crc kubenswrapper[4696]: I1125 12:00:30.802234 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:00:30 crc kubenswrapper[4696]: I1125 12:00:30.802677 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:00:38 crc kubenswrapper[4696]: I1125 12:00:38.013580 4696 scope.go:117] "RemoveContainer" containerID="cfe2671f84e41ed45fd6b0e3b68559864f3c5f3b7fbd7793c5170a517a8e334d" Nov 25 12:00:38 crc kubenswrapper[4696]: I1125 12:00:38.062970 4696 scope.go:117] "RemoveContainer" containerID="2ba619c864c03a483e99401b0e6f0bb0c9161bbbe0a67c3d79e27c68860e30ca" Nov 25 12:00:38 crc kubenswrapper[4696]: I1125 12:00:38.112236 4696 scope.go:117] "RemoveContainer" containerID="1ff2982196ab48c6d88d9e20c3deeac311bd2b393a6f40d008d3501d635d57f5" Nov 25 12:00:38 crc kubenswrapper[4696]: I1125 12:00:38.137243 4696 scope.go:117] "RemoveContainer" containerID="a6ec6bbe22b7d3bd32175ed73434ccc0a279bef928f201474d6ecdf388702a1e" Nov 25 12:00:38 crc kubenswrapper[4696]: I1125 12:00:38.197971 4696 scope.go:117] "RemoveContainer" containerID="6a58d1fa30495bdc32e09ac77dc6c1189b63e258dffe865ac90df83bf8ead2dd" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.160124 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401201-9gntv"] Nov 25 12:01:00 crc kubenswrapper[4696]: E1125 12:01:00.161249 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680e9982-47c3-43b3-ab5e-4d83de22aaea" containerName="collect-profiles" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.161269 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="680e9982-47c3-43b3-ab5e-4d83de22aaea" containerName="collect-profiles" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.161524 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="680e9982-47c3-43b3-ab5e-4d83de22aaea" containerName="collect-profiles" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.162366 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.187092 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401201-9gntv"] Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.332632 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-config-data\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.333054 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-fernet-keys\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.333086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-combined-ca-bundle\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.333479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwjq4\" (UniqueName: \"kubernetes.io/projected/fe1b84e6-203d-46c1-9681-de4c6e8badef-kube-api-access-mwjq4\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.435642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwjq4\" (UniqueName: \"kubernetes.io/projected/fe1b84e6-203d-46c1-9681-de4c6e8badef-kube-api-access-mwjq4\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.435794 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-config-data\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.435840 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-fernet-keys\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.435862 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-combined-ca-bundle\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.442441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-config-data\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.442468 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-fernet-keys\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.447719 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-combined-ca-bundle\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.461405 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwjq4\" (UniqueName: \"kubernetes.io/projected/fe1b84e6-203d-46c1-9681-de4c6e8badef-kube-api-access-mwjq4\") pod \"keystone-cron-29401201-9gntv\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.502095 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.805322 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:01:00 crc kubenswrapper[4696]: I1125 12:01:00.805580 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:01:01 crc kubenswrapper[4696]: I1125 12:01:01.229299 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401201-9gntv"] Nov 25 12:01:01 crc kubenswrapper[4696]: I1125 12:01:01.786847 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-9gntv" event={"ID":"fe1b84e6-203d-46c1-9681-de4c6e8badef","Type":"ContainerStarted","Data":"fcaaef0359b4c77bf98a5a03465dd0bd138e335de2e25486d8f9d6cd5a32dc7d"} Nov 25 12:01:01 crc kubenswrapper[4696]: I1125 12:01:01.787232 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-9gntv" event={"ID":"fe1b84e6-203d-46c1-9681-de4c6e8badef","Type":"ContainerStarted","Data":"0ed1ab5cfcbf0426d9fa7f1b1d0c4bf87943d5b82ec455b19e9e712faf54897e"} Nov 25 12:01:01 crc kubenswrapper[4696]: I1125 12:01:01.806036 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401201-9gntv" podStartSLOduration=1.806013325 podStartE2EDuration="1.806013325s" podCreationTimestamp="2025-11-25 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:01:01.803252748 +0000 UTC m=+5058.616869345" watchObservedRunningTime="2025-11-25 12:01:01.806013325 +0000 UTC m=+5058.619629912" Nov 25 12:01:06 crc kubenswrapper[4696]: I1125 12:01:06.838998 4696 generic.go:334] "Generic (PLEG): container finished" podID="fe1b84e6-203d-46c1-9681-de4c6e8badef" containerID="fcaaef0359b4c77bf98a5a03465dd0bd138e335de2e25486d8f9d6cd5a32dc7d" exitCode=0 Nov 25 12:01:06 crc kubenswrapper[4696]: I1125 12:01:06.839076 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-9gntv" event={"ID":"fe1b84e6-203d-46c1-9681-de4c6e8badef","Type":"ContainerDied","Data":"fcaaef0359b4c77bf98a5a03465dd0bd138e335de2e25486d8f9d6cd5a32dc7d"} Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.224893 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.409261 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-combined-ca-bundle\") pod \"fe1b84e6-203d-46c1-9681-de4c6e8badef\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.409737 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwjq4\" (UniqueName: \"kubernetes.io/projected/fe1b84e6-203d-46c1-9681-de4c6e8badef-kube-api-access-mwjq4\") pod \"fe1b84e6-203d-46c1-9681-de4c6e8badef\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.409946 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-fernet-keys\") pod \"fe1b84e6-203d-46c1-9681-de4c6e8badef\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.409975 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-config-data\") pod \"fe1b84e6-203d-46c1-9681-de4c6e8badef\" (UID: \"fe1b84e6-203d-46c1-9681-de4c6e8badef\") " Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.415023 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1b84e6-203d-46c1-9681-de4c6e8badef-kube-api-access-mwjq4" (OuterVolumeSpecName: "kube-api-access-mwjq4") pod "fe1b84e6-203d-46c1-9681-de4c6e8badef" (UID: "fe1b84e6-203d-46c1-9681-de4c6e8badef"). InnerVolumeSpecName "kube-api-access-mwjq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.416365 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fe1b84e6-203d-46c1-9681-de4c6e8badef" (UID: "fe1b84e6-203d-46c1-9681-de4c6e8badef"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.440262 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe1b84e6-203d-46c1-9681-de4c6e8badef" (UID: "fe1b84e6-203d-46c1-9681-de4c6e8badef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.472938 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-config-data" (OuterVolumeSpecName: "config-data") pod "fe1b84e6-203d-46c1-9681-de4c6e8badef" (UID: "fe1b84e6-203d-46c1-9681-de4c6e8badef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.512315 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.512348 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.512357 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b84e6-203d-46c1-9681-de4c6e8badef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.512369 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwjq4\" (UniqueName: \"kubernetes.io/projected/fe1b84e6-203d-46c1-9681-de4c6e8badef-kube-api-access-mwjq4\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.860584 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-9gntv" event={"ID":"fe1b84e6-203d-46c1-9681-de4c6e8badef","Type":"ContainerDied","Data":"0ed1ab5cfcbf0426d9fa7f1b1d0c4bf87943d5b82ec455b19e9e712faf54897e"} Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.860870 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ed1ab5cfcbf0426d9fa7f1b1d0c4bf87943d5b82ec455b19e9e712faf54897e" Nov 25 12:01:08 crc kubenswrapper[4696]: I1125 12:01:08.860632 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-9gntv" Nov 25 12:01:30 crc kubenswrapper[4696]: I1125 12:01:30.801945 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:01:30 crc kubenswrapper[4696]: I1125 12:01:30.802566 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:01:30 crc kubenswrapper[4696]: I1125 12:01:30.802621 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 12:01:30 crc kubenswrapper[4696]: I1125 12:01:30.803407 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"332ab86e2b889db1ce953dd2bf5d3716c0deb9acbbf4dfb647276bf2b6d5b2b2"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:01:30 crc kubenswrapper[4696]: I1125 12:01:30.803465 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://332ab86e2b889db1ce953dd2bf5d3716c0deb9acbbf4dfb647276bf2b6d5b2b2" gracePeriod=600 Nov 25 12:01:31 crc kubenswrapper[4696]: I1125 12:01:31.087464 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="332ab86e2b889db1ce953dd2bf5d3716c0deb9acbbf4dfb647276bf2b6d5b2b2" exitCode=0 Nov 25 12:01:31 crc kubenswrapper[4696]: I1125 12:01:31.087559 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"332ab86e2b889db1ce953dd2bf5d3716c0deb9acbbf4dfb647276bf2b6d5b2b2"} Nov 25 12:01:31 crc kubenswrapper[4696]: I1125 12:01:31.087846 4696 scope.go:117] "RemoveContainer" containerID="fbeb69ea706619543ed016b43757ec9cf3dee4afad4f9b040fbddf5910bcb59b" Nov 25 12:01:32 crc kubenswrapper[4696]: I1125 12:01:32.098575 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1"} Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.292166 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-j8vkp/must-gather-ztcs8"] Nov 25 12:02:34 crc kubenswrapper[4696]: E1125 12:02:34.293100 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1b84e6-203d-46c1-9681-de4c6e8badef" containerName="keystone-cron" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.293118 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1b84e6-203d-46c1-9681-de4c6e8badef" containerName="keystone-cron" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.293350 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1b84e6-203d-46c1-9681-de4c6e8badef" containerName="keystone-cron" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.294564 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.297653 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-j8vkp"/"openshift-service-ca.crt" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.297673 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-j8vkp"/"kube-root-ca.crt" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.325876 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-j8vkp/must-gather-ztcs8"] Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.463346 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16cc0183-a87c-4fab-92cb-35f2e7570194-must-gather-output\") pod \"must-gather-ztcs8\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.463792 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlzsf\" (UniqueName: \"kubernetes.io/projected/16cc0183-a87c-4fab-92cb-35f2e7570194-kube-api-access-nlzsf\") pod \"must-gather-ztcs8\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.566072 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16cc0183-a87c-4fab-92cb-35f2e7570194-must-gather-output\") pod \"must-gather-ztcs8\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.566153 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlzsf\" (UniqueName: \"kubernetes.io/projected/16cc0183-a87c-4fab-92cb-35f2e7570194-kube-api-access-nlzsf\") pod \"must-gather-ztcs8\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.566441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16cc0183-a87c-4fab-92cb-35f2e7570194-must-gather-output\") pod \"must-gather-ztcs8\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.586411 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlzsf\" (UniqueName: \"kubernetes.io/projected/16cc0183-a87c-4fab-92cb-35f2e7570194-kube-api-access-nlzsf\") pod \"must-gather-ztcs8\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:34 crc kubenswrapper[4696]: I1125 12:02:34.614538 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:02:35 crc kubenswrapper[4696]: I1125 12:02:35.201462 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-j8vkp/must-gather-ztcs8"] Nov 25 12:02:35 crc kubenswrapper[4696]: I1125 12:02:35.709133 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" event={"ID":"16cc0183-a87c-4fab-92cb-35f2e7570194","Type":"ContainerStarted","Data":"2aa5402a3978f65d31e9bfa120e1bd71c174a7e4a6b638e91042aacbaaa3a8dd"} Nov 25 12:02:36 crc kubenswrapper[4696]: I1125 12:02:36.725535 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" event={"ID":"16cc0183-a87c-4fab-92cb-35f2e7570194","Type":"ContainerStarted","Data":"4dae7561a5266c1bfa848994bbf2ed293d84f5e77cbf4543717ac62eadfeb023"} Nov 25 12:02:36 crc kubenswrapper[4696]: I1125 12:02:36.726109 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" event={"ID":"16cc0183-a87c-4fab-92cb-35f2e7570194","Type":"ContainerStarted","Data":"4c1efc41bbf359ff7a4b7763d77266714366da4d7944dc489985f892c75c079f"} Nov 25 12:02:36 crc kubenswrapper[4696]: I1125 12:02:36.753984 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" podStartSLOduration=2.753959509 podStartE2EDuration="2.753959509s" podCreationTimestamp="2025-11-25 12:02:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:02:36.751497391 +0000 UTC m=+5153.565113988" watchObservedRunningTime="2025-11-25 12:02:36.753959509 +0000 UTC m=+5153.567576096" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.546893 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-6zqjv"] Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.550053 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.553080 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-j8vkp"/"default-dockercfg-gps5p" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.728590 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltq2r\" (UniqueName: \"kubernetes.io/projected/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-kube-api-access-ltq2r\") pod \"crc-debug-6zqjv\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.728953 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-host\") pod \"crc-debug-6zqjv\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.831191 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltq2r\" (UniqueName: \"kubernetes.io/projected/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-kube-api-access-ltq2r\") pod \"crc-debug-6zqjv\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.831326 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-host\") pod \"crc-debug-6zqjv\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.831473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-host\") pod \"crc-debug-6zqjv\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.864005 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltq2r\" (UniqueName: \"kubernetes.io/projected/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-kube-api-access-ltq2r\") pod \"crc-debug-6zqjv\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:40 crc kubenswrapper[4696]: I1125 12:02:40.870335 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:02:41 crc kubenswrapper[4696]: I1125 12:02:41.778382 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" event={"ID":"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc","Type":"ContainerStarted","Data":"1075a45d422742176dc2b49f0a650153ec4ddc11f6f8c968b0ae12ee52d72266"} Nov 25 12:02:41 crc kubenswrapper[4696]: I1125 12:02:41.779064 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" event={"ID":"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc","Type":"ContainerStarted","Data":"1afbda31c6dfb20fe950fc7f5786385c64df430e1f15529f9365de2d4d312b1e"} Nov 25 12:02:41 crc kubenswrapper[4696]: I1125 12:02:41.796514 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" podStartSLOduration=1.796493365 podStartE2EDuration="1.796493365s" podCreationTimestamp="2025-11-25 12:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:02:41.793438328 +0000 UTC m=+5158.607054925" watchObservedRunningTime="2025-11-25 12:02:41.796493365 +0000 UTC m=+5158.610109952" Nov 25 12:03:33 crc kubenswrapper[4696]: I1125 12:03:33.305097 4696 generic.go:334] "Generic (PLEG): container finished" podID="c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" containerID="1075a45d422742176dc2b49f0a650153ec4ddc11f6f8c968b0ae12ee52d72266" exitCode=0 Nov 25 12:03:33 crc kubenswrapper[4696]: I1125 12:03:33.305163 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" event={"ID":"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc","Type":"ContainerDied","Data":"1075a45d422742176dc2b49f0a650153ec4ddc11f6f8c968b0ae12ee52d72266"} Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.452068 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.537302 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-6zqjv"] Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.571677 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-6zqjv"] Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.590711 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltq2r\" (UniqueName: \"kubernetes.io/projected/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-kube-api-access-ltq2r\") pod \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.590863 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-host\") pod \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\" (UID: \"c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc\") " Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.590937 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-host" (OuterVolumeSpecName: "host") pod "c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" (UID: "c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.591430 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-host\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.608108 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-kube-api-access-ltq2r" (OuterVolumeSpecName: "kube-api-access-ltq2r") pod "c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" (UID: "c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc"). InnerVolumeSpecName "kube-api-access-ltq2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:03:34 crc kubenswrapper[4696]: I1125 12:03:34.693159 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltq2r\" (UniqueName: \"kubernetes.io/projected/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc-kube-api-access-ltq2r\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.330632 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1afbda31c6dfb20fe950fc7f5786385c64df430e1f15529f9365de2d4d312b1e" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.330882 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-6zqjv" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.690313 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-l6pcg"] Nov 25 12:03:35 crc kubenswrapper[4696]: E1125 12:03:35.690805 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" containerName="container-00" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.690823 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" containerName="container-00" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.691088 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" containerName="container-00" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.691876 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.694874 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-j8vkp"/"default-dockercfg-gps5p" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.815446 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/63d3cfbf-8624-4705-b7e8-23a52601ca47-host\") pod \"crc-debug-l6pcg\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.815804 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6nnn\" (UniqueName: \"kubernetes.io/projected/63d3cfbf-8624-4705-b7e8-23a52601ca47-kube-api-access-h6nnn\") pod \"crc-debug-l6pcg\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.918296 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6nnn\" (UniqueName: \"kubernetes.io/projected/63d3cfbf-8624-4705-b7e8-23a52601ca47-kube-api-access-h6nnn\") pod \"crc-debug-l6pcg\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.918819 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/63d3cfbf-8624-4705-b7e8-23a52601ca47-host\") pod \"crc-debug-l6pcg\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.918916 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/63d3cfbf-8624-4705-b7e8-23a52601ca47-host\") pod \"crc-debug-l6pcg\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:35 crc kubenswrapper[4696]: I1125 12:03:35.941231 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6nnn\" (UniqueName: \"kubernetes.io/projected/63d3cfbf-8624-4705-b7e8-23a52601ca47-kube-api-access-h6nnn\") pod \"crc-debug-l6pcg\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:36 crc kubenswrapper[4696]: I1125 12:03:36.024299 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:36 crc kubenswrapper[4696]: I1125 12:03:36.059113 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc" path="/var/lib/kubelet/pods/c7e03ff5-8073-4ffa-9cf3-cdec6e90d3fc/volumes" Nov 25 12:03:36 crc kubenswrapper[4696]: I1125 12:03:36.341883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" event={"ID":"63d3cfbf-8624-4705-b7e8-23a52601ca47","Type":"ContainerStarted","Data":"a39ef180d808945f26ba82eb72d9eb3c26999b5f732ce054b05b8dc9c76045de"} Nov 25 12:03:37 crc kubenswrapper[4696]: I1125 12:03:37.353698 4696 generic.go:334] "Generic (PLEG): container finished" podID="63d3cfbf-8624-4705-b7e8-23a52601ca47" containerID="44c80f26246c538d0ed71345d5f80163ed11682fb7476ada5db84557b10e9e47" exitCode=0 Nov 25 12:03:37 crc kubenswrapper[4696]: I1125 12:03:37.353843 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" event={"ID":"63d3cfbf-8624-4705-b7e8-23a52601ca47","Type":"ContainerDied","Data":"44c80f26246c538d0ed71345d5f80163ed11682fb7476ada5db84557b10e9e47"} Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.521116 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.684685 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6nnn\" (UniqueName: \"kubernetes.io/projected/63d3cfbf-8624-4705-b7e8-23a52601ca47-kube-api-access-h6nnn\") pod \"63d3cfbf-8624-4705-b7e8-23a52601ca47\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.686019 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/63d3cfbf-8624-4705-b7e8-23a52601ca47-host\") pod \"63d3cfbf-8624-4705-b7e8-23a52601ca47\" (UID: \"63d3cfbf-8624-4705-b7e8-23a52601ca47\") " Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.686180 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63d3cfbf-8624-4705-b7e8-23a52601ca47-host" (OuterVolumeSpecName: "host") pod "63d3cfbf-8624-4705-b7e8-23a52601ca47" (UID: "63d3cfbf-8624-4705-b7e8-23a52601ca47"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.688914 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/63d3cfbf-8624-4705-b7e8-23a52601ca47-host\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.719015 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d3cfbf-8624-4705-b7e8-23a52601ca47-kube-api-access-h6nnn" (OuterVolumeSpecName: "kube-api-access-h6nnn") pod "63d3cfbf-8624-4705-b7e8-23a52601ca47" (UID: "63d3cfbf-8624-4705-b7e8-23a52601ca47"). InnerVolumeSpecName "kube-api-access-h6nnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:03:38 crc kubenswrapper[4696]: I1125 12:03:38.792406 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6nnn\" (UniqueName: \"kubernetes.io/projected/63d3cfbf-8624-4705-b7e8-23a52601ca47-kube-api-access-h6nnn\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:39 crc kubenswrapper[4696]: I1125 12:03:39.381808 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" event={"ID":"63d3cfbf-8624-4705-b7e8-23a52601ca47","Type":"ContainerDied","Data":"a39ef180d808945f26ba82eb72d9eb3c26999b5f732ce054b05b8dc9c76045de"} Nov 25 12:03:39 crc kubenswrapper[4696]: I1125 12:03:39.382252 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a39ef180d808945f26ba82eb72d9eb3c26999b5f732ce054b05b8dc9c76045de" Nov 25 12:03:39 crc kubenswrapper[4696]: I1125 12:03:39.382259 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-l6pcg" Nov 25 12:03:39 crc kubenswrapper[4696]: I1125 12:03:39.744161 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-l6pcg"] Nov 25 12:03:39 crc kubenswrapper[4696]: I1125 12:03:39.754220 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-l6pcg"] Nov 25 12:03:40 crc kubenswrapper[4696]: I1125 12:03:40.053895 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d3cfbf-8624-4705-b7e8-23a52601ca47" path="/var/lib/kubelet/pods/63d3cfbf-8624-4705-b7e8-23a52601ca47/volumes" Nov 25 12:03:40 crc kubenswrapper[4696]: I1125 12:03:40.972660 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-pl4rh"] Nov 25 12:03:40 crc kubenswrapper[4696]: E1125 12:03:40.973480 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d3cfbf-8624-4705-b7e8-23a52601ca47" containerName="container-00" Nov 25 12:03:40 crc kubenswrapper[4696]: I1125 12:03:40.973497 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d3cfbf-8624-4705-b7e8-23a52601ca47" containerName="container-00" Nov 25 12:03:40 crc kubenswrapper[4696]: I1125 12:03:40.973704 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d3cfbf-8624-4705-b7e8-23a52601ca47" containerName="container-00" Nov 25 12:03:40 crc kubenswrapper[4696]: I1125 12:03:40.974332 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:40 crc kubenswrapper[4696]: I1125 12:03:40.977256 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-j8vkp"/"default-dockercfg-gps5p" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.137457 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7v56\" (UniqueName: \"kubernetes.io/projected/09b91168-95e5-402f-9a09-7dec2528b152-kube-api-access-d7v56\") pod \"crc-debug-pl4rh\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.137562 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09b91168-95e5-402f-9a09-7dec2528b152-host\") pod \"crc-debug-pl4rh\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.239781 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09b91168-95e5-402f-9a09-7dec2528b152-host\") pod \"crc-debug-pl4rh\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.240001 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7v56\" (UniqueName: \"kubernetes.io/projected/09b91168-95e5-402f-9a09-7dec2528b152-kube-api-access-d7v56\") pod \"crc-debug-pl4rh\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.241763 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09b91168-95e5-402f-9a09-7dec2528b152-host\") pod \"crc-debug-pl4rh\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.271871 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7v56\" (UniqueName: \"kubernetes.io/projected/09b91168-95e5-402f-9a09-7dec2528b152-kube-api-access-d7v56\") pod \"crc-debug-pl4rh\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.296373 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:41 crc kubenswrapper[4696]: W1125 12:03:41.363606 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b WatchSource:0}: Error finding container 7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b: Status 404 returned error can't find the container with id 7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b Nov 25 12:03:41 crc kubenswrapper[4696]: I1125 12:03:41.405053 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" event={"ID":"09b91168-95e5-402f-9a09-7dec2528b152","Type":"ContainerStarted","Data":"7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b"} Nov 25 12:03:42 crc kubenswrapper[4696]: I1125 12:03:42.418811 4696 generic.go:334] "Generic (PLEG): container finished" podID="09b91168-95e5-402f-9a09-7dec2528b152" containerID="272b2bbd6c4903675ae7859e50f3a7e3d92a0fdd863bcbcb81e8f1e4e36cb852" exitCode=0 Nov 25 12:03:42 crc kubenswrapper[4696]: I1125 12:03:42.418925 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" event={"ID":"09b91168-95e5-402f-9a09-7dec2528b152","Type":"ContainerDied","Data":"272b2bbd6c4903675ae7859e50f3a7e3d92a0fdd863bcbcb81e8f1e4e36cb852"} Nov 25 12:03:42 crc kubenswrapper[4696]: I1125 12:03:42.461524 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-pl4rh"] Nov 25 12:03:42 crc kubenswrapper[4696]: I1125 12:03:42.473838 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-j8vkp/crc-debug-pl4rh"] Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.532987 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.705351 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7v56\" (UniqueName: \"kubernetes.io/projected/09b91168-95e5-402f-9a09-7dec2528b152-kube-api-access-d7v56\") pod \"09b91168-95e5-402f-9a09-7dec2528b152\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.705780 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09b91168-95e5-402f-9a09-7dec2528b152-host\") pod \"09b91168-95e5-402f-9a09-7dec2528b152\" (UID: \"09b91168-95e5-402f-9a09-7dec2528b152\") " Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.706241 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09b91168-95e5-402f-9a09-7dec2528b152-host" (OuterVolumeSpecName: "host") pod "09b91168-95e5-402f-9a09-7dec2528b152" (UID: "09b91168-95e5-402f-9a09-7dec2528b152"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.712426 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b91168-95e5-402f-9a09-7dec2528b152-kube-api-access-d7v56" (OuterVolumeSpecName: "kube-api-access-d7v56") pod "09b91168-95e5-402f-9a09-7dec2528b152" (UID: "09b91168-95e5-402f-9a09-7dec2528b152"). InnerVolumeSpecName "kube-api-access-d7v56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.808126 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7v56\" (UniqueName: \"kubernetes.io/projected/09b91168-95e5-402f-9a09-7dec2528b152-kube-api-access-d7v56\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:43 crc kubenswrapper[4696]: I1125 12:03:43.808546 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09b91168-95e5-402f-9a09-7dec2528b152-host\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:44 crc kubenswrapper[4696]: I1125 12:03:44.058980 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b91168-95e5-402f-9a09-7dec2528b152" path="/var/lib/kubelet/pods/09b91168-95e5-402f-9a09-7dec2528b152/volumes" Nov 25 12:03:44 crc kubenswrapper[4696]: I1125 12:03:44.439499 4696 scope.go:117] "RemoveContainer" containerID="272b2bbd6c4903675ae7859e50f3a7e3d92a0fdd863bcbcb81e8f1e4e36cb852" Nov 25 12:03:44 crc kubenswrapper[4696]: I1125 12:03:44.439631 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/crc-debug-pl4rh" Nov 25 12:03:45 crc kubenswrapper[4696]: E1125 12:03:45.295126 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b\": RecentStats: unable to find data in memory cache]" Nov 25 12:03:55 crc kubenswrapper[4696]: E1125 12:03:55.628126 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice\": RecentStats: unable to find data in memory cache]" Nov 25 12:04:00 crc kubenswrapper[4696]: I1125 12:04:00.801692 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:04:00 crc kubenswrapper[4696]: I1125 12:04:00.802376 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:04:05 crc kubenswrapper[4696]: E1125 12:04:05.960327 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b\": RecentStats: unable to find data in memory cache]" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.201359 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wm9c4"] Nov 25 12:04:12 crc kubenswrapper[4696]: E1125 12:04:12.202857 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b91168-95e5-402f-9a09-7dec2528b152" containerName="container-00" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.202876 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b91168-95e5-402f-9a09-7dec2528b152" containerName="container-00" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.203114 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b91168-95e5-402f-9a09-7dec2528b152" containerName="container-00" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.205352 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.213319 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wm9c4"] Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.298134 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-utilities\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.298233 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-catalog-content\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.298333 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsjw9\" (UniqueName: \"kubernetes.io/projected/13d11d67-e798-4b94-be9f-d7bb916065a0-kube-api-access-dsjw9\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.400303 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-catalog-content\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.400373 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsjw9\" (UniqueName: \"kubernetes.io/projected/13d11d67-e798-4b94-be9f-d7bb916065a0-kube-api-access-dsjw9\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.400481 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-utilities\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.400822 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-catalog-content\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.401161 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-utilities\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.425043 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsjw9\" (UniqueName: \"kubernetes.io/projected/13d11d67-e798-4b94-be9f-d7bb916065a0-kube-api-access-dsjw9\") pod \"redhat-operators-wm9c4\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:12 crc kubenswrapper[4696]: I1125 12:04:12.528028 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:13 crc kubenswrapper[4696]: I1125 12:04:13.140724 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wm9c4"] Nov 25 12:04:13 crc kubenswrapper[4696]: W1125 12:04:13.157348 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13d11d67_e798_4b94_be9f_d7bb916065a0.slice/crio-f8a94f6e42e96e0cdd9ec5e7bc512e803d4905c5c59d92a7f4117eefcb9311e0 WatchSource:0}: Error finding container f8a94f6e42e96e0cdd9ec5e7bc512e803d4905c5c59d92a7f4117eefcb9311e0: Status 404 returned error can't find the container with id f8a94f6e42e96e0cdd9ec5e7bc512e803d4905c5c59d92a7f4117eefcb9311e0 Nov 25 12:04:13 crc kubenswrapper[4696]: I1125 12:04:13.725396 4696 generic.go:334] "Generic (PLEG): container finished" podID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerID="c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890" exitCode=0 Nov 25 12:04:13 crc kubenswrapper[4696]: I1125 12:04:13.725525 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerDied","Data":"c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890"} Nov 25 12:04:13 crc kubenswrapper[4696]: I1125 12:04:13.725772 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerStarted","Data":"f8a94f6e42e96e0cdd9ec5e7bc512e803d4905c5c59d92a7f4117eefcb9311e0"} Nov 25 12:04:13 crc kubenswrapper[4696]: I1125 12:04:13.728265 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:04:15 crc kubenswrapper[4696]: I1125 12:04:15.748794 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerStarted","Data":"982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b"} Nov 25 12:04:16 crc kubenswrapper[4696]: E1125 12:04:16.240948 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b\": RecentStats: unable to find data in memory cache]" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.773514 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9fjtk"] Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.779005 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.802635 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fjtk"] Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.894329 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-utilities\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.894401 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-catalog-content\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.894478 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4krcc\" (UniqueName: \"kubernetes.io/projected/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-kube-api-access-4krcc\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.995823 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-catalog-content\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.995942 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4krcc\" (UniqueName: \"kubernetes.io/projected/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-kube-api-access-4krcc\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.996044 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-utilities\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.996461 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-catalog-content\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:16 crc kubenswrapper[4696]: I1125 12:04:16.996506 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-utilities\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:17 crc kubenswrapper[4696]: I1125 12:04:17.018249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4krcc\" (UniqueName: \"kubernetes.io/projected/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-kube-api-access-4krcc\") pod \"community-operators-9fjtk\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:17 crc kubenswrapper[4696]: I1125 12:04:17.099885 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:17 crc kubenswrapper[4696]: I1125 12:04:17.548707 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fjtk"] Nov 25 12:04:17 crc kubenswrapper[4696]: W1125 12:04:17.555291 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cbc8d0c_14cc_41e9_bb7f_18aefbc8071c.slice/crio-aa8c55cfc3ad87af17986bdced460c73d019a7608bb4e9b0f9a3d29e08ab18b2 WatchSource:0}: Error finding container aa8c55cfc3ad87af17986bdced460c73d019a7608bb4e9b0f9a3d29e08ab18b2: Status 404 returned error can't find the container with id aa8c55cfc3ad87af17986bdced460c73d019a7608bb4e9b0f9a3d29e08ab18b2 Nov 25 12:04:17 crc kubenswrapper[4696]: I1125 12:04:17.772124 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerStarted","Data":"aa8c55cfc3ad87af17986bdced460c73d019a7608bb4e9b0f9a3d29e08ab18b2"} Nov 25 12:04:18 crc kubenswrapper[4696]: I1125 12:04:18.782431 4696 generic.go:334] "Generic (PLEG): container finished" podID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerID="091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884" exitCode=0 Nov 25 12:04:18 crc kubenswrapper[4696]: I1125 12:04:18.782545 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerDied","Data":"091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884"} Nov 25 12:04:24 crc kubenswrapper[4696]: I1125 12:04:24.842015 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerStarted","Data":"4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485"} Nov 25 12:04:26 crc kubenswrapper[4696]: E1125 12:04:26.544900 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice\": RecentStats: unable to find data in memory cache]" Nov 25 12:04:28 crc kubenswrapper[4696]: I1125 12:04:28.879825 4696 generic.go:334] "Generic (PLEG): container finished" podID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerID="982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b" exitCode=0 Nov 25 12:04:28 crc kubenswrapper[4696]: I1125 12:04:28.879877 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerDied","Data":"982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b"} Nov 25 12:04:29 crc kubenswrapper[4696]: I1125 12:04:29.892730 4696 generic.go:334] "Generic (PLEG): container finished" podID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerID="4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485" exitCode=0 Nov 25 12:04:29 crc kubenswrapper[4696]: I1125 12:04:29.892815 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerDied","Data":"4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485"} Nov 25 12:04:29 crc kubenswrapper[4696]: I1125 12:04:29.897179 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerStarted","Data":"ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2"} Nov 25 12:04:29 crc kubenswrapper[4696]: I1125 12:04:29.934026 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wm9c4" podStartSLOduration=2.325889536 podStartE2EDuration="17.934005707s" podCreationTimestamp="2025-11-25 12:04:12 +0000 UTC" firstStartedPulling="2025-11-25 12:04:13.728051171 +0000 UTC m=+5250.541667758" lastFinishedPulling="2025-11-25 12:04:29.336167352 +0000 UTC m=+5266.149783929" observedRunningTime="2025-11-25 12:04:29.931741275 +0000 UTC m=+5266.745357872" watchObservedRunningTime="2025-11-25 12:04:29.934005707 +0000 UTC m=+5266.747622294" Nov 25 12:04:30 crc kubenswrapper[4696]: I1125 12:04:30.802423 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:04:30 crc kubenswrapper[4696]: I1125 12:04:30.802986 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:04:30 crc kubenswrapper[4696]: I1125 12:04:30.913881 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerStarted","Data":"ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8"} Nov 25 12:04:30 crc kubenswrapper[4696]: I1125 12:04:30.964380 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9fjtk" podStartSLOduration=3.411841457 podStartE2EDuration="14.964354507s" podCreationTimestamp="2025-11-25 12:04:16 +0000 UTC" firstStartedPulling="2025-11-25 12:04:18.784704934 +0000 UTC m=+5255.598321521" lastFinishedPulling="2025-11-25 12:04:30.337217984 +0000 UTC m=+5267.150834571" observedRunningTime="2025-11-25 12:04:30.957869283 +0000 UTC m=+5267.771485870" watchObservedRunningTime="2025-11-25 12:04:30.964354507 +0000 UTC m=+5267.777971094" Nov 25 12:04:31 crc kubenswrapper[4696]: I1125 12:04:31.924564 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9c4d748fd-g4t2c_c0f652ee-e64b-423b-902a-bab2bd12d46f/barbican-api/0.log" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.058352 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9c4d748fd-g4t2c_c0f652ee-e64b-423b-902a-bab2bd12d46f/barbican-api-log/0.log" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.262325 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-66cc7b58db-xph79_9e0f5310-cf60-4125-9b91-2dfda5383c7d/barbican-keystone-listener/0.log" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.339886 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-66cc7b58db-xph79_9e0f5310-cf60-4125-9b91-2dfda5383c7d/barbican-keystone-listener-log/0.log" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.528278 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.529514 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.596037 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64b77754f5-27gpf_22dbae09-113b-4136-b9d6-6162815b5ef7/barbican-worker/0.log" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.709971 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64b77754f5-27gpf_22dbae09-113b-4136-b9d6-6162815b5ef7/barbican-worker-log/0.log" Nov 25 12:04:32 crc kubenswrapper[4696]: I1125 12:04:32.906713 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-65952_062b8e49-6137-4b89-8ae4-a29272ff0d1f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.123334 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/ceilometer-notification-agent/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.161300 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/proxy-httpd/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.184564 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/ceilometer-central-agent/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.310775 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9fa6192b-7ea2-414b-9e4f-ebb5648a58bf/sg-core/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.411296 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5a8e77cd-7f3d-45e6-bb36-2ce208f70adb/cinder-api-log/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.529422 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5a8e77cd-7f3d-45e6-bb36-2ce208f70adb/cinder-api/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.608119 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wm9c4" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" probeResult="failure" output=< Nov 25 12:04:33 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 12:04:33 crc kubenswrapper[4696]: > Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.799923 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_dccb9177-738d-433c-a2ec-99131e690f89/cinder-scheduler/0.log" Nov 25 12:04:33 crc kubenswrapper[4696]: I1125 12:04:33.916776 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_dccb9177-738d-433c-a2ec-99131e690f89/probe/0.log" Nov 25 12:04:34 crc kubenswrapper[4696]: I1125 12:04:34.000573 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-qr5kp_574f645b-d7d9-4756-82ef-27bfa7bf982e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:34 crc kubenswrapper[4696]: I1125 12:04:34.274764 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bzgww_114399b2-d253-44d9-994e-5098ef72a972/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:34 crc kubenswrapper[4696]: I1125 12:04:34.378979 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-qdbhs_6b74d08c-9100-49d2-8483-34d4c76c0c6b/init/0.log" Nov 25 12:04:34 crc kubenswrapper[4696]: I1125 12:04:34.766245 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-qdbhs_6b74d08c-9100-49d2-8483-34d4c76c0c6b/init/0.log" Nov 25 12:04:34 crc kubenswrapper[4696]: I1125 12:04:34.877800 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-j29t5_38409cac-958b-497e-9431-4d0cafe2fd8d/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:35 crc kubenswrapper[4696]: I1125 12:04:35.060415 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-qdbhs_6b74d08c-9100-49d2-8483-34d4c76c0c6b/dnsmasq-dns/0.log" Nov 25 12:04:35 crc kubenswrapper[4696]: I1125 12:04:35.327848 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d642739c-282a-4644-9aae-68f82c3ddf58/glance-httpd/0.log" Nov 25 12:04:35 crc kubenswrapper[4696]: I1125 12:04:35.348600 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d642739c-282a-4644-9aae-68f82c3ddf58/glance-log/0.log" Nov 25 12:04:35 crc kubenswrapper[4696]: I1125 12:04:35.920117 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_bd3e6e15-4a89-468d-8151-26c9e4c19dc5/glance-httpd/0.log" Nov 25 12:04:36 crc kubenswrapper[4696]: I1125 12:04:36.040648 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_bd3e6e15-4a89-468d-8151-26c9e4c19dc5/glance-log/0.log" Nov 25 12:04:36 crc kubenswrapper[4696]: I1125 12:04:36.235489 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6558fc797b-2qp4r_1944527b-6ffc-4b6b-954f-4b01394ea0cf/horizon/1.log" Nov 25 12:04:36 crc kubenswrapper[4696]: I1125 12:04:36.401649 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6558fc797b-2qp4r_1944527b-6ffc-4b6b-954f-4b01394ea0cf/horizon/0.log" Nov 25 12:04:36 crc kubenswrapper[4696]: I1125 12:04:36.756743 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-zxfw6_3202ae74-b34c-41aa-b717-6313d9e71bf4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:36 crc kubenswrapper[4696]: E1125 12:04:36.915468 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice/crio-7e568c5bc15c8f58f64cdaa969a5bd8c5348b0c2c068e4da232d36fe0d39f68b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b91168_95e5_402f_9a09_7dec2528b152.slice\": RecentStats: unable to find data in memory cache]" Nov 25 12:04:37 crc kubenswrapper[4696]: I1125 12:04:37.104874 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:37 crc kubenswrapper[4696]: I1125 12:04:37.105368 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:37 crc kubenswrapper[4696]: I1125 12:04:37.215466 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kg6v5_d69df5c8-e28a-4b73-9d2b-97087aaed351/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:37 crc kubenswrapper[4696]: I1125 12:04:37.234265 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6558fc797b-2qp4r_1944527b-6ffc-4b6b-954f-4b01394ea0cf/horizon-log/0.log" Nov 25 12:04:37 crc kubenswrapper[4696]: I1125 12:04:37.686834 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401141-q8gxc_0a3da9a0-6262-4ff8-b510-e476db9d9985/keystone-cron/0.log" Nov 25 12:04:37 crc kubenswrapper[4696]: I1125 12:04:37.777602 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5547bcb8fc-b65gm_ad61fdbd-b908-4430-8b43-e08260f0dbd1/keystone-api/0.log" Nov 25 12:04:38 crc kubenswrapper[4696]: I1125 12:04:38.009844 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401201-9gntv_fe1b84e6-203d-46c1-9681-de4c6e8badef/keystone-cron/0.log" Nov 25 12:04:38 crc kubenswrapper[4696]: I1125 12:04:38.019956 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9ed1a610-3fb4-4643-a481-724f5036b836/kube-state-metrics/0.log" Nov 25 12:04:38 crc kubenswrapper[4696]: I1125 12:04:38.177202 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-9fjtk" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="registry-server" probeResult="failure" output=< Nov 25 12:04:38 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 12:04:38 crc kubenswrapper[4696]: > Nov 25 12:04:38 crc kubenswrapper[4696]: I1125 12:04:38.343226 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-gm7jr_27d6547e-1c82-4b39-a281-476463f32e81/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:39 crc kubenswrapper[4696]: I1125 12:04:39.225986 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c599bcb65-hgc2v_412cdec0-a9b1-4796-82f6-95c1f6b9b6ae/neutron-httpd/0.log" Nov 25 12:04:39 crc kubenswrapper[4696]: I1125 12:04:39.291557 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-5crns_9813095c-910d-451b-b3bb-18a6d5fcdf26/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:39 crc kubenswrapper[4696]: I1125 12:04:39.311243 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c599bcb65-hgc2v_412cdec0-a9b1-4796-82f6-95c1f6b9b6ae/neutron-api/0.log" Nov 25 12:04:40 crc kubenswrapper[4696]: I1125 12:04:40.586367 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8189ed6c-e99d-44db-bd26-7b74ea6539fb/nova-cell0-conductor-conductor/0.log" Nov 25 12:04:40 crc kubenswrapper[4696]: I1125 12:04:40.883989 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_984434ed-24d1-4090-a7d9-c38b39617b74/nova-cell1-conductor-conductor/0.log" Nov 25 12:04:41 crc kubenswrapper[4696]: I1125 12:04:41.473867 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_672502a7-5ef3-4596-9507-21afa27d44e6/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 12:04:41 crc kubenswrapper[4696]: I1125 12:04:41.617421 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8377dca6-a47c-4ffb-9916-f6505eed1e48/nova-api-log/0.log" Nov 25 12:04:41 crc kubenswrapper[4696]: I1125 12:04:41.775329 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-4fkn4_11902ad3-7187-4749-8fbf-3d5f1eb77764/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:42 crc kubenswrapper[4696]: I1125 12:04:42.178290 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8377dca6-a47c-4ffb-9916-f6505eed1e48/nova-api-api/0.log" Nov 25 12:04:42 crc kubenswrapper[4696]: I1125 12:04:42.193499 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d808ab02-7a75-4441-a5d0-b92fcaa1eb06/nova-metadata-log/0.log" Nov 25 12:04:42 crc kubenswrapper[4696]: I1125 12:04:42.901467 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c541fdf-4691-4f61-9865-8497d636667d/mysql-bootstrap/0.log" Nov 25 12:04:43 crc kubenswrapper[4696]: I1125 12:04:43.234525 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c541fdf-4691-4f61-9865-8497d636667d/mysql-bootstrap/0.log" Nov 25 12:04:43 crc kubenswrapper[4696]: I1125 12:04:43.253818 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c541fdf-4691-4f61-9865-8497d636667d/galera/0.log" Nov 25 12:04:43 crc kubenswrapper[4696]: I1125 12:04:43.461091 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_97905b5e-c4ba-4712-8ed6-e1f1c228f683/nova-scheduler-scheduler/0.log" Nov 25 12:04:43 crc kubenswrapper[4696]: I1125 12:04:43.596051 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ad9b8866-501c-48e8-a694-5fff74e631fc/mysql-bootstrap/0.log" Nov 25 12:04:43 crc kubenswrapper[4696]: I1125 12:04:43.663266 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wm9c4" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" probeResult="failure" output=< Nov 25 12:04:43 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 12:04:43 crc kubenswrapper[4696]: > Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.004336 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ad9b8866-501c-48e8-a694-5fff74e631fc/mysql-bootstrap/0.log" Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.033793 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ad9b8866-501c-48e8-a694-5fff74e631fc/galera/0.log" Nov 25 12:04:44 crc kubenswrapper[4696]: E1125 12:04:44.092213 4696 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/4074a6c2eebc2ebcbf7b0d30009438e6f9d3f474b11986af92e49bb02cca0e05/diff" to get inode usage: stat /var/lib/containers/storage/overlay/4074a6c2eebc2ebcbf7b0d30009438e6f9d3f474b11986af92e49bb02cca0e05/diff: no such file or directory, extraDiskErr: Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.424050 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jz4h7_87dc2400-74ac-4b0b-ae7d-ca62010572d3/ovn-controller/0.log" Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.429267 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_14829ad2-5154-4bc3-ac0d-8343c93c0f6d/openstackclient/0.log" Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.755598 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_eba40268-f540-4ac7-9ae4-14455c81395f/memcached/0.log" Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.785756 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-qld29_d900f0ea-8197-4383-bdbb-666b60db9f19/openstack-network-exporter/0.log" Nov 25 12:04:44 crc kubenswrapper[4696]: I1125 12:04:44.876793 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d808ab02-7a75-4441-a5d0-b92fcaa1eb06/nova-metadata-metadata/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.051735 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovsdb-server-init/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.272962 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovsdb-server-init/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.278372 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovsdb-server/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.301306 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7w2gw_cb7e6448-16df-4635-84d9-97f6ee770c36/ovs-vswitchd/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.431658 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-zf8h7_348a4b45-dfe7-449f-8b03-871394bfba46/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.607572 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e0d794d3-7682-4248-affe-b53f4f34f264/ovn-northd/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.729563 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e0d794d3-7682-4248-affe-b53f4f34f264/openstack-network-exporter/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.752051 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8b4e55fc-7bf4-4a72-b805-b7b2776de33c/openstack-network-exporter/0.log" Nov 25 12:04:45 crc kubenswrapper[4696]: I1125 12:04:45.907219 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8b4e55fc-7bf4-4a72-b805-b7b2776de33c/ovsdbserver-nb/0.log" Nov 25 12:04:46 crc kubenswrapper[4696]: I1125 12:04:46.321616 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbc1467c-cf49-44a9-99eb-c592c38412a1/openstack-network-exporter/0.log" Nov 25 12:04:46 crc kubenswrapper[4696]: I1125 12:04:46.403368 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbc1467c-cf49-44a9-99eb-c592c38412a1/ovsdbserver-sb/0.log" Nov 25 12:04:46 crc kubenswrapper[4696]: I1125 12:04:46.716790 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-646fccfb54-pbhr5_537410db-3589-4c1a-bb28-1f0a66a20ff6/placement-api/0.log" Nov 25 12:04:46 crc kubenswrapper[4696]: I1125 12:04:46.748523 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf5b7831-caa9-4411-9386-cf99ba7f8a20/setup-container/0.log" Nov 25 12:04:46 crc kubenswrapper[4696]: I1125 12:04:46.870363 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-646fccfb54-pbhr5_537410db-3589-4c1a-bb28-1f0a66a20ff6/placement-log/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.183867 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf5b7831-caa9-4411-9386-cf99ba7f8a20/rabbitmq/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.187114 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.228253 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b161351-5bdf-4cfe-895b-29381797d908/setup-container/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.265805 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.271788 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf5b7831-caa9-4411-9386-cf99ba7f8a20/setup-container/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.551471 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b161351-5bdf-4cfe-895b-29381797d908/rabbitmq/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.554708 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b161351-5bdf-4cfe-895b-29381797d908/setup-container/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.680971 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-glmr7_5779201b-7609-43c1-9cec-c3cc15302542/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.894798 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ls6nj_2a97a4bb-2255-4f00-901a-3392d7808c53/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:47 crc kubenswrapper[4696]: I1125 12:04:47.984036 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fjtk"] Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.035825 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-hhmzx_4fbc74d0-e08c-4e4e-a05f-b07bf9cc5b1a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.139147 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-w8rsq_438e2932-b4ff-4135-b4a8-7b4e89d574fc/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.313708 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9w2md_d97e4985-25b9-4106-8053-b7ee1aa42542/ssh-known-hosts-edpm-deployment/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.431895 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78446968f5-84c89_142ffa20-cd2a-4b6b-90ff-2770e6eed020/proxy-server/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.557023 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78446968f5-84c89_142ffa20-cd2a-4b6b-90ff-2770e6eed020/proxy-httpd/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.747616 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-xshf5_dbeda82d-3956-47f7-95a8-6ffc25961e5d/swift-ring-rebalance/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.774546 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-auditor/0.log" Nov 25 12:04:48 crc kubenswrapper[4696]: I1125 12:04:48.874485 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-reaper/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.045882 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-auditor/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.113596 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9fjtk" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="registry-server" containerID="cri-o://ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8" gracePeriod=2 Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.152213 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-server/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.180108 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/account-replicator/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.431317 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-replicator/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.612149 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-server/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.760069 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/container-updater/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.922067 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-auditor/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.934253 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-expirer/0.log" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.949563 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:49 crc kubenswrapper[4696]: I1125 12:04:49.969185 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-replicator/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.100256 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-catalog-content\") pod \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.100445 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4krcc\" (UniqueName: \"kubernetes.io/projected/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-kube-api-access-4krcc\") pod \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.100580 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-utilities\") pod \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\" (UID: \"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c\") " Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.101950 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-utilities" (OuterVolumeSpecName: "utilities") pod "5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" (UID: "5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.106052 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-updater/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.107944 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-kube-api-access-4krcc" (OuterVolumeSpecName: "kube-api-access-4krcc") pod "5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" (UID: "5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c"). InnerVolumeSpecName "kube-api-access-4krcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.130705 4696 generic.go:334] "Generic (PLEG): container finished" podID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerID="ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8" exitCode=0 Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.130761 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerDied","Data":"ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8"} Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.130797 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fjtk" event={"ID":"5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c","Type":"ContainerDied","Data":"aa8c55cfc3ad87af17986bdced460c73d019a7608bb4e9b0f9a3d29e08ab18b2"} Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.130822 4696 scope.go:117] "RemoveContainer" containerID="ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.130881 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fjtk" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.136009 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/object-server/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.173510 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" (UID: "5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.174094 4696 scope.go:117] "RemoveContainer" containerID="4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.202888 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.202936 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.202951 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4krcc\" (UniqueName: \"kubernetes.io/projected/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c-kube-api-access-4krcc\") on node \"crc\" DevicePath \"\"" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.205048 4696 scope.go:117] "RemoveContainer" containerID="091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.273191 4696 scope.go:117] "RemoveContainer" containerID="ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8" Nov 25 12:04:50 crc kubenswrapper[4696]: E1125 12:04:50.277861 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8\": container with ID starting with ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8 not found: ID does not exist" containerID="ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.277913 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8"} err="failed to get container status \"ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8\": rpc error: code = NotFound desc = could not find container \"ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8\": container with ID starting with ff4fc5a841f82f70788e2d924a67da3e02aa4181e58853e8f00924a428e689a8 not found: ID does not exist" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.277944 4696 scope.go:117] "RemoveContainer" containerID="4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.278040 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/rsync/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: E1125 12:04:50.284064 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485\": container with ID starting with 4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485 not found: ID does not exist" containerID="4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.284118 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485"} err="failed to get container status \"4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485\": rpc error: code = NotFound desc = could not find container \"4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485\": container with ID starting with 4335657020f9fc18a2ae307e97cf56b100b0e49dabc45a2fb24b6aaa35f75485 not found: ID does not exist" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.284159 4696 scope.go:117] "RemoveContainer" containerID="091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884" Nov 25 12:04:50 crc kubenswrapper[4696]: E1125 12:04:50.288565 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884\": container with ID starting with 091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884 not found: ID does not exist" containerID="091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.288655 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884"} err="failed to get container status \"091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884\": rpc error: code = NotFound desc = could not find container \"091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884\": container with ID starting with 091ca7908ac6b4025a327433fdbc1eba019fa0b10319f9adacc1d763604aa884 not found: ID does not exist" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.381350 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e87af5be-8499-4303-a158-8c9d50c2db73/swift-recon-cron/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.468857 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fjtk"] Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.480658 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9fjtk"] Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.583894 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-pzkxz_372198a7-8311-4133-8047-fcb9908b7f90/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.713817 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_f8d03f21-fad7-432e-8220-12094b55a1a1/tempest-tests-tempest-tests-runner/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.717641 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5df67e6a-3bdf-4da9-9599-0d7e1d69e6b6/test-operator-logs-container/0.log" Nov 25 12:04:50 crc kubenswrapper[4696]: I1125 12:04:50.917925 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7lc7s_b756844d-cdfd-4345-8a32-035b71161bfc/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 12:04:52 crc kubenswrapper[4696]: I1125 12:04:52.055814 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" path="/var/lib/kubelet/pods/5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c/volumes" Nov 25 12:04:53 crc kubenswrapper[4696]: I1125 12:04:53.577426 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wm9c4" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" probeResult="failure" output=< Nov 25 12:04:53 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 12:04:53 crc kubenswrapper[4696]: > Nov 25 12:05:00 crc kubenswrapper[4696]: I1125 12:05:00.802214 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:05:00 crc kubenswrapper[4696]: I1125 12:05:00.804012 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:05:00 crc kubenswrapper[4696]: I1125 12:05:00.804156 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" Nov 25 12:05:00 crc kubenswrapper[4696]: I1125 12:05:00.805111 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1"} pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:05:00 crc kubenswrapper[4696]: I1125 12:05:00.805312 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" containerID="cri-o://0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" gracePeriod=600 Nov 25 12:05:00 crc kubenswrapper[4696]: E1125 12:05:00.939775 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:05:01 crc kubenswrapper[4696]: I1125 12:05:01.234517 4696 generic.go:334] "Generic (PLEG): container finished" podID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" exitCode=0 Nov 25 12:05:01 crc kubenswrapper[4696]: I1125 12:05:01.234569 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerDied","Data":"0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1"} Nov 25 12:05:01 crc kubenswrapper[4696]: I1125 12:05:01.234609 4696 scope.go:117] "RemoveContainer" containerID="332ab86e2b889db1ce953dd2bf5d3716c0deb9acbbf4dfb647276bf2b6d5b2b2" Nov 25 12:05:01 crc kubenswrapper[4696]: I1125 12:05:01.235406 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:05:01 crc kubenswrapper[4696]: E1125 12:05:01.235756 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:05:03 crc kubenswrapper[4696]: I1125 12:05:03.598492 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wm9c4" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" probeResult="failure" output=< Nov 25 12:05:03 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Nov 25 12:05:03 crc kubenswrapper[4696]: > Nov 25 12:05:12 crc kubenswrapper[4696]: I1125 12:05:12.597019 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:05:12 crc kubenswrapper[4696]: I1125 12:05:12.671304 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:05:13 crc kubenswrapper[4696]: I1125 12:05:13.445633 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wm9c4"] Nov 25 12:05:14 crc kubenswrapper[4696]: I1125 12:05:14.363267 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wm9c4" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" containerID="cri-o://ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2" gracePeriod=2 Nov 25 12:05:14 crc kubenswrapper[4696]: I1125 12:05:14.906639 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.095517 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-utilities\") pod \"13d11d67-e798-4b94-be9f-d7bb916065a0\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.096010 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-catalog-content\") pod \"13d11d67-e798-4b94-be9f-d7bb916065a0\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.096146 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsjw9\" (UniqueName: \"kubernetes.io/projected/13d11d67-e798-4b94-be9f-d7bb916065a0-kube-api-access-dsjw9\") pod \"13d11d67-e798-4b94-be9f-d7bb916065a0\" (UID: \"13d11d67-e798-4b94-be9f-d7bb916065a0\") " Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.096418 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-utilities" (OuterVolumeSpecName: "utilities") pod "13d11d67-e798-4b94-be9f-d7bb916065a0" (UID: "13d11d67-e798-4b94-be9f-d7bb916065a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.096740 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.104619 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13d11d67-e798-4b94-be9f-d7bb916065a0-kube-api-access-dsjw9" (OuterVolumeSpecName: "kube-api-access-dsjw9") pod "13d11d67-e798-4b94-be9f-d7bb916065a0" (UID: "13d11d67-e798-4b94-be9f-d7bb916065a0"). InnerVolumeSpecName "kube-api-access-dsjw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.198825 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsjw9\" (UniqueName: \"kubernetes.io/projected/13d11d67-e798-4b94-be9f-d7bb916065a0-kube-api-access-dsjw9\") on node \"crc\" DevicePath \"\"" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.210985 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13d11d67-e798-4b94-be9f-d7bb916065a0" (UID: "13d11d67-e798-4b94-be9f-d7bb916065a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.300941 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d11d67-e798-4b94-be9f-d7bb916065a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.379941 4696 generic.go:334] "Generic (PLEG): container finished" podID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerID="ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2" exitCode=0 Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.379993 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerDied","Data":"ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2"} Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.380024 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm9c4" event={"ID":"13d11d67-e798-4b94-be9f-d7bb916065a0","Type":"ContainerDied","Data":"f8a94f6e42e96e0cdd9ec5e7bc512e803d4905c5c59d92a7f4117eefcb9311e0"} Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.380042 4696 scope.go:117] "RemoveContainer" containerID="ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.380207 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm9c4" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.408844 4696 scope.go:117] "RemoveContainer" containerID="982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.426775 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wm9c4"] Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.437725 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wm9c4"] Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.442891 4696 scope.go:117] "RemoveContainer" containerID="c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.492857 4696 scope.go:117] "RemoveContainer" containerID="ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2" Nov 25 12:05:15 crc kubenswrapper[4696]: E1125 12:05:15.493428 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2\": container with ID starting with ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2 not found: ID does not exist" containerID="ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.493475 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2"} err="failed to get container status \"ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2\": rpc error: code = NotFound desc = could not find container \"ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2\": container with ID starting with ca7644a7946e585d423e33167c6fd33cc4be8a4e54473f3ebbbe48ff0e301ba2 not found: ID does not exist" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.493501 4696 scope.go:117] "RemoveContainer" containerID="982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b" Nov 25 12:05:15 crc kubenswrapper[4696]: E1125 12:05:15.493940 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b\": container with ID starting with 982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b not found: ID does not exist" containerID="982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.494000 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b"} err="failed to get container status \"982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b\": rpc error: code = NotFound desc = could not find container \"982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b\": container with ID starting with 982250accf370936e42f44788b80baba7bec30de32c17c94c6b8620a5cb35d7b not found: ID does not exist" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.494035 4696 scope.go:117] "RemoveContainer" containerID="c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890" Nov 25 12:05:15 crc kubenswrapper[4696]: E1125 12:05:15.494394 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890\": container with ID starting with c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890 not found: ID does not exist" containerID="c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890" Nov 25 12:05:15 crc kubenswrapper[4696]: I1125 12:05:15.494426 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890"} err="failed to get container status \"c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890\": rpc error: code = NotFound desc = could not find container \"c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890\": container with ID starting with c526d70ddee9421ba042c3dbcaccac66ebdee1ade4e906a79670e9e206694890 not found: ID does not exist" Nov 25 12:05:16 crc kubenswrapper[4696]: I1125 12:05:16.042883 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:05:16 crc kubenswrapper[4696]: E1125 12:05:16.043872 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:05:16 crc kubenswrapper[4696]: I1125 12:05:16.056107 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" path="/var/lib/kubelet/pods/13d11d67-e798-4b94-be9f-d7bb916065a0/volumes" Nov 25 12:05:21 crc kubenswrapper[4696]: I1125 12:05:21.915898 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/util/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.132892 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/pull/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.151543 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/util/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.157562 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/pull/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.384628 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/pull/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.395647 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/extract/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.404437 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3e3367b5c270c6e014d30a7a230c0f6d017d403995104a44ddc21ea7204jbl5_e0a75471-aabe-4787-a3ec-fedcf4cecbf1/util/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.611845 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-4kpwn_c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd/kube-rbac-proxy/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.673049 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-4kpwn_c83f0c59-a2fa-49a7-99bf-9e3ff30f1bfd/manager/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.786518 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-5kdhx_900bb9ad-0f16-4eb4-beb5-c23e0089db39/kube-rbac-proxy/0.log" Nov 25 12:05:22 crc kubenswrapper[4696]: I1125 12:05:22.997644 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-vrf9f_2ae50512-8525-4878-80b1-7e65a4c035d9/kube-rbac-proxy/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.040110 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-5kdhx_900bb9ad-0f16-4eb4-beb5-c23e0089db39/manager/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.075645 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-vrf9f_2ae50512-8525-4878-80b1-7e65a4c035d9/manager/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.305893 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-ldvdx_7844a3a7-9ef3-48b7-9ab8-944304690567/kube-rbac-proxy/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.393462 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-ldvdx_7844a3a7-9ef3-48b7-9ab8-944304690567/manager/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.520204 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-c7l6c_28a8de6d-a7e1-4ca9-853d-6d27a066df12/manager/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.590149 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-c7l6c_28a8de6d-a7e1-4ca9-853d-6d27a066df12/kube-rbac-proxy/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.670493 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2p89s_347bfeb5-b62b-4515-be16-1fdb0fc52e44/kube-rbac-proxy/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.800396 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2p89s_347bfeb5-b62b-4515-be16-1fdb0fc52e44/manager/0.log" Nov 25 12:05:23 crc kubenswrapper[4696]: I1125 12:05:23.865970 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-2xtrg_f8dc76df-0e30-435e-b61e-926c2b60a87e/kube-rbac-proxy/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.101653 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-2xtrg_f8dc76df-0e30-435e-b61e-926c2b60a87e/manager/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.107548 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q4ml6_a891b8d2-92ed-4f76-9947-3968ee474921/kube-rbac-proxy/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.234253 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-q4ml6_a891b8d2-92ed-4f76-9947-3968ee474921/manager/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.392034 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-4fkts_01eb2add-76e2-414f-8a70-275d2a1b8939/manager/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.431203 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-4fkts_01eb2add-76e2-414f-8a70-275d2a1b8939/kube-rbac-proxy/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.633121 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-xwzk7_ce501661-3cfd-4cce-8256-c347638da2f6/manager/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.714610 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-xwzk7_ce501661-3cfd-4cce-8256-c347638da2f6/kube-rbac-proxy/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.814836 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-vdjlx_394e9c9d-8d79-4be4-9679-432c658bd52d/kube-rbac-proxy/0.log" Nov 25 12:05:24 crc kubenswrapper[4696]: I1125 12:05:24.960843 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-vdjlx_394e9c9d-8d79-4be4-9679-432c658bd52d/manager/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.085118 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-xqqtv_8faec67a-3309-496b-9c5a-74eb6421a7f2/kube-rbac-proxy/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.129176 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-xqqtv_8faec67a-3309-496b-9c5a-74eb6421a7f2/manager/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.308628 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-7xgqt_c40524c0-d668-4b1c-8113-67b152b150e7/kube-rbac-proxy/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.499613 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-7xgqt_c40524c0-d668-4b1c-8113-67b152b150e7/manager/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.532133 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-gjl4t_c6ca26a1-b284-48b2-b68c-fd0c83b404ec/kube-rbac-proxy/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.690831 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-gjl4t_c6ca26a1-b284-48b2-b68c-fd0c83b404ec/manager/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.774964 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t_e6871bcf-31f0-4782-b94c-56e1002cd0d1/kube-rbac-proxy/0.log" Nov 25 12:05:25 crc kubenswrapper[4696]: I1125 12:05:25.896046 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vvz6t_e6871bcf-31f0-4782-b94c-56e1002cd0d1/manager/0.log" Nov 25 12:05:26 crc kubenswrapper[4696]: I1125 12:05:26.385253 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cccc9d798-j4nfn_62e71f26-481d-4d1f-b3c7-bed2f91968cd/operator/0.log" Nov 25 12:05:26 crc kubenswrapper[4696]: I1125 12:05:26.535431 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-55w9h_2d7b9c54-45bc-4bc8-9513-a37f199b022f/registry-server/0.log" Nov 25 12:05:26 crc kubenswrapper[4696]: I1125 12:05:26.729579 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-599kv_7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f/kube-rbac-proxy/0.log" Nov 25 12:05:26 crc kubenswrapper[4696]: I1125 12:05:26.849960 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-599kv_7b7e7428-ccde-4fdb-b9f9-a9d971d4d46f/manager/0.log" Nov 25 12:05:26 crc kubenswrapper[4696]: I1125 12:05:26.993039 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-gnbb8_33329628-2334-4d45-ad9f-20d2d23ab8ba/kube-rbac-proxy/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.042036 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:05:27 crc kubenswrapper[4696]: E1125 12:05:27.042464 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.098296 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-gnbb8_33329628-2334-4d45-ad9f-20d2d23ab8ba/manager/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.419388 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-9ln5q_33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2/kube-rbac-proxy/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.478179 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65d98ff956-rrwtq_ad6baa72-3698-484a-a2f2-690719d71677/manager/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.510538 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94k5m_3367886a-6027-418f-9415-38d775304079/operator/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.580867 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-9ln5q_33f71b0b-dfb5-4cb5-a8bc-1dda8d515cd2/manager/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.695331 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-jfrk4_90aa0c65-64c1-4ec3-a104-122691247673/kube-rbac-proxy/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.858380 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-jfrk4_90aa0c65-64c1-4ec3-a104-122691247673/manager/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.871715 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-q2c8n_3f468970-3367-47b7-82fa-e9af018c3996/kube-rbac-proxy/0.log" Nov 25 12:05:27 crc kubenswrapper[4696]: I1125 12:05:27.980474 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-q2c8n_3f468970-3367-47b7-82fa-e9af018c3996/manager/0.log" Nov 25 12:05:28 crc kubenswrapper[4696]: I1125 12:05:28.039771 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-758cd_17643fa7-c097-4a6c-ac8a-fe7ec0cc9901/kube-rbac-proxy/0.log" Nov 25 12:05:28 crc kubenswrapper[4696]: I1125 12:05:28.099336 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-758cd_17643fa7-c097-4a6c-ac8a-fe7ec0cc9901/manager/0.log" Nov 25 12:05:41 crc kubenswrapper[4696]: I1125 12:05:41.042530 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:05:41 crc kubenswrapper[4696]: E1125 12:05:41.043377 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:05:48 crc kubenswrapper[4696]: I1125 12:05:48.384624 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-n62qw_2f6f8fdc-fec0-46b9-9277-2996da698f24/control-plane-machine-set-operator/0.log" Nov 25 12:05:48 crc kubenswrapper[4696]: I1125 12:05:48.685830 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5g9hp_23a01a08-d008-451f-86c0-f3c456fafb88/kube-rbac-proxy/0.log" Nov 25 12:05:48 crc kubenswrapper[4696]: I1125 12:05:48.727623 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5g9hp_23a01a08-d008-451f-86c0-f3c456fafb88/machine-api-operator/0.log" Nov 25 12:05:56 crc kubenswrapper[4696]: I1125 12:05:56.042479 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:05:56 crc kubenswrapper[4696]: E1125 12:05:56.044495 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:06:01 crc kubenswrapper[4696]: I1125 12:06:01.934052 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-dvr5m_c4daf7cb-c6fd-43da-8821-d26fe9f88fac/cert-manager-controller/0.log" Nov 25 12:06:02 crc kubenswrapper[4696]: I1125 12:06:02.173508 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-knqmh_ed104e46-ddf7-496c-a389-1c78d24e8e7c/cert-manager-webhook/0.log" Nov 25 12:06:02 crc kubenswrapper[4696]: I1125 12:06:02.192233 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qt985_c1f44937-600c-4df9-94f4-8adbc7a9b0ef/cert-manager-cainjector/0.log" Nov 25 12:06:07 crc kubenswrapper[4696]: I1125 12:06:07.043035 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:06:07 crc kubenswrapper[4696]: E1125 12:06:07.043843 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:06:15 crc kubenswrapper[4696]: I1125 12:06:15.920187 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-52cq5_b35b7a4c-2b0c-47ed-8c2f-3cb5a5d11332/nmstate-console-plugin/0.log" Nov 25 12:06:16 crc kubenswrapper[4696]: I1125 12:06:16.081137 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ttnxf_a2a61ba7-423c-4fbc-9bf5-63af41ee1170/nmstate-handler/0.log" Nov 25 12:06:16 crc kubenswrapper[4696]: I1125 12:06:16.151207 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-xznfr_585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d/nmstate-metrics/0.log" Nov 25 12:06:16 crc kubenswrapper[4696]: I1125 12:06:16.246740 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-xznfr_585d3d73-0434-4ce9-b2e7-04bdaf2e7c1d/kube-rbac-proxy/0.log" Nov 25 12:06:16 crc kubenswrapper[4696]: I1125 12:06:16.383015 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-wkpl2_d246ee4c-dd53-42a6-b24f-5e34687b6c8e/nmstate-operator/0.log" Nov 25 12:06:16 crc kubenswrapper[4696]: I1125 12:06:16.558458 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-5bpk8_fc74e56a-3f49-4d38-b5e4-78cafb3276a6/nmstate-webhook/0.log" Nov 25 12:06:21 crc kubenswrapper[4696]: I1125 12:06:21.041900 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:06:21 crc kubenswrapper[4696]: E1125 12:06:21.042620 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:06:31 crc kubenswrapper[4696]: I1125 12:06:31.927165 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-jgmgn_b24625b7-df3d-4f50-ba97-5eac956b1662/kube-rbac-proxy/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.000331 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-jgmgn_b24625b7-df3d-4f50-ba97-5eac956b1662/controller/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.149472 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.305016 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.338219 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.421503 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.462088 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.633222 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.714808 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.714846 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.751309 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 12:06:32 crc kubenswrapper[4696]: I1125 12:06:32.966375 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-reloader/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.014810 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-metrics/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.015364 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/cp-frr-files/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.019642 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/controller/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.191654 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/frr-metrics/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.206078 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/kube-rbac-proxy/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.337878 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/kube-rbac-proxy-frr/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.473484 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/reloader/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.594066 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-nd8g6_c2b796fe-37ae-4c76-b1a0-fe8e916f8c6f/frr-k8s-webhook-server/0.log" Nov 25 12:06:33 crc kubenswrapper[4696]: I1125 12:06:33.945288 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d5845fdbd-84v4v_5ad9cb9e-8e80-41bc-bb0a-687f580830b0/manager/0.log" Nov 25 12:06:34 crc kubenswrapper[4696]: I1125 12:06:34.113298 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b88d4b6cb-m8sz5_c8010a11-0748-4b8b-b40d-14d57fe53f40/webhook-server/0.log" Nov 25 12:06:34 crc kubenswrapper[4696]: I1125 12:06:34.287860 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42dc9_e4999207-9162-4635-b7a8-2bd0d5311019/kube-rbac-proxy/0.log" Nov 25 12:06:34 crc kubenswrapper[4696]: I1125 12:06:34.980640 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42dc9_e4999207-9162-4635-b7a8-2bd0d5311019/speaker/0.log" Nov 25 12:06:35 crc kubenswrapper[4696]: I1125 12:06:35.026397 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhrzx_1462b48a-8222-418b-bcb1-06b9d3a2f99b/frr/0.log" Nov 25 12:06:36 crc kubenswrapper[4696]: I1125 12:06:36.042374 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:06:36 crc kubenswrapper[4696]: E1125 12:06:36.043763 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:06:47 crc kubenswrapper[4696]: I1125 12:06:47.522371 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/util/0.log" Nov 25 12:06:47 crc kubenswrapper[4696]: I1125 12:06:47.769227 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/util/0.log" Nov 25 12:06:47 crc kubenswrapper[4696]: I1125 12:06:47.772752 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/pull/0.log" Nov 25 12:06:47 crc kubenswrapper[4696]: I1125 12:06:47.778939 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/pull/0.log" Nov 25 12:06:47 crc kubenswrapper[4696]: I1125 12:06:47.992719 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/util/0.log" Nov 25 12:06:47 crc kubenswrapper[4696]: I1125 12:06:47.996798 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/pull/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.005870 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772evqxqn_7cce94e7-26a7-4e28-8880-7a47b2d5116f/extract/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.188259 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-utilities/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.370916 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-content/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.386208 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-utilities/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.459490 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-content/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.580426 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-content/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.623220 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/extract-utilities/0.log" Nov 25 12:06:48 crc kubenswrapper[4696]: I1125 12:06:48.848892 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-utilities/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.077172 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9hlrx_25e551df-fa3d-40bd-b2e0-e972976f5c74/registry-server/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.180456 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-content/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.180853 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-utilities/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.232294 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-content/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.392683 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-content/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.472895 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/extract-utilities/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.749025 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/util/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.923004 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/pull/0.log" Nov 25 12:06:49 crc kubenswrapper[4696]: I1125 12:06:49.928988 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/util/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.051590 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/pull/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.349609 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/pull/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.366941 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/util/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.432435 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbc7f_4ba6a060-a9da-44c9-b1c8-450ea6075e18/registry-server/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.439389 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6w9m72_ec4528a0-140d-4ae5-b1dd-c3f8fc133ef4/extract/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.637929 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xrn8c_e77b3a6e-c1e9-44ef-806f-6adcba39fe31/marketplace-operator/1.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.714537 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xrn8c_e77b3a6e-c1e9-44ef-806f-6adcba39fe31/marketplace-operator/0.log" Nov 25 12:06:50 crc kubenswrapper[4696]: I1125 12:06:50.884260 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-utilities/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.042830 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:06:51 crc kubenswrapper[4696]: E1125 12:06:51.043045 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.060842 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-content/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.106110 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-content/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.142953 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-utilities/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.289153 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-utilities/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.309700 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/extract-content/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.592291 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-utilities/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.632511 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h9ssz_df2ddc89-6af7-4142-a3bd-58856bf7eb77/registry-server/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.820859 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-utilities/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.877074 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-content/0.log" Nov 25 12:06:51 crc kubenswrapper[4696]: I1125 12:06:51.889605 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-content/0.log" Nov 25 12:06:52 crc kubenswrapper[4696]: I1125 12:06:52.104900 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-content/0.log" Nov 25 12:06:52 crc kubenswrapper[4696]: I1125 12:06:52.124777 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/extract-utilities/0.log" Nov 25 12:06:52 crc kubenswrapper[4696]: I1125 12:06:52.248633 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-llfmt_ff7ff079-c37c-4973-852c-3c6afbadcda5/registry-server/0.log" Nov 25 12:07:05 crc kubenswrapper[4696]: I1125 12:07:05.043308 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:07:05 crc kubenswrapper[4696]: E1125 12:07:05.044072 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:07:19 crc kubenswrapper[4696]: I1125 12:07:19.042067 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:07:19 crc kubenswrapper[4696]: E1125 12:07:19.042690 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:07:34 crc kubenswrapper[4696]: I1125 12:07:34.062565 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:07:34 crc kubenswrapper[4696]: E1125 12:07:34.063817 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:07:48 crc kubenswrapper[4696]: I1125 12:07:48.042420 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:07:48 crc kubenswrapper[4696]: E1125 12:07:48.043242 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:08:00 crc kubenswrapper[4696]: I1125 12:08:00.042654 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:08:00 crc kubenswrapper[4696]: E1125 12:08:00.043475 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:08:15 crc kubenswrapper[4696]: I1125 12:08:15.042164 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:08:15 crc kubenswrapper[4696]: E1125 12:08:15.042952 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:08:28 crc kubenswrapper[4696]: I1125 12:08:28.042170 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:08:28 crc kubenswrapper[4696]: E1125 12:08:28.043048 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:08:40 crc kubenswrapper[4696]: I1125 12:08:40.043192 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:08:40 crc kubenswrapper[4696]: E1125 12:08:40.044069 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:08:52 crc kubenswrapper[4696]: I1125 12:08:52.042865 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:08:52 crc kubenswrapper[4696]: E1125 12:08:52.043767 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:09:06 crc kubenswrapper[4696]: I1125 12:09:06.042845 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:09:06 crc kubenswrapper[4696]: E1125 12:09:06.045090 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:09:17 crc kubenswrapper[4696]: I1125 12:09:17.042415 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:09:17 crc kubenswrapper[4696]: E1125 12:09:17.043288 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:09:18 crc kubenswrapper[4696]: I1125 12:09:18.622692 4696 generic.go:334] "Generic (PLEG): container finished" podID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerID="4c1efc41bbf359ff7a4b7763d77266714366da4d7944dc489985f892c75c079f" exitCode=0 Nov 25 12:09:18 crc kubenswrapper[4696]: I1125 12:09:18.622782 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" event={"ID":"16cc0183-a87c-4fab-92cb-35f2e7570194","Type":"ContainerDied","Data":"4c1efc41bbf359ff7a4b7763d77266714366da4d7944dc489985f892c75c079f"} Nov 25 12:09:18 crc kubenswrapper[4696]: I1125 12:09:18.623555 4696 scope.go:117] "RemoveContainer" containerID="4c1efc41bbf359ff7a4b7763d77266714366da4d7944dc489985f892c75c079f" Nov 25 12:09:19 crc kubenswrapper[4696]: I1125 12:09:19.146956 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j8vkp_must-gather-ztcs8_16cc0183-a87c-4fab-92cb-35f2e7570194/gather/0.log" Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.044186 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:09:32 crc kubenswrapper[4696]: E1125 12:09:32.045580 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.505848 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-j8vkp/must-gather-ztcs8"] Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.506386 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="copy" containerID="cri-o://4dae7561a5266c1bfa848994bbf2ed293d84f5e77cbf4543717ac62eadfeb023" gracePeriod=2 Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.516695 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-j8vkp/must-gather-ztcs8"] Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.781989 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j8vkp_must-gather-ztcs8_16cc0183-a87c-4fab-92cb-35f2e7570194/copy/0.log" Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.787112 4696 generic.go:334] "Generic (PLEG): container finished" podID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerID="4dae7561a5266c1bfa848994bbf2ed293d84f5e77cbf4543717ac62eadfeb023" exitCode=143 Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.951261 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j8vkp_must-gather-ztcs8_16cc0183-a87c-4fab-92cb-35f2e7570194/copy/0.log" Nov 25 12:09:32 crc kubenswrapper[4696]: I1125 12:09:32.952042 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.060685 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlzsf\" (UniqueName: \"kubernetes.io/projected/16cc0183-a87c-4fab-92cb-35f2e7570194-kube-api-access-nlzsf\") pod \"16cc0183-a87c-4fab-92cb-35f2e7570194\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.060814 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16cc0183-a87c-4fab-92cb-35f2e7570194-must-gather-output\") pod \"16cc0183-a87c-4fab-92cb-35f2e7570194\" (UID: \"16cc0183-a87c-4fab-92cb-35f2e7570194\") " Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.068444 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16cc0183-a87c-4fab-92cb-35f2e7570194-kube-api-access-nlzsf" (OuterVolumeSpecName: "kube-api-access-nlzsf") pod "16cc0183-a87c-4fab-92cb-35f2e7570194" (UID: "16cc0183-a87c-4fab-92cb-35f2e7570194"). InnerVolumeSpecName "kube-api-access-nlzsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.167546 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlzsf\" (UniqueName: \"kubernetes.io/projected/16cc0183-a87c-4fab-92cb-35f2e7570194-kube-api-access-nlzsf\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.263810 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16cc0183-a87c-4fab-92cb-35f2e7570194-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "16cc0183-a87c-4fab-92cb-35f2e7570194" (UID: "16cc0183-a87c-4fab-92cb-35f2e7570194"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.268274 4696 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16cc0183-a87c-4fab-92cb-35f2e7570194-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.817742 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j8vkp_must-gather-ztcs8_16cc0183-a87c-4fab-92cb-35f2e7570194/copy/0.log" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.818837 4696 scope.go:117] "RemoveContainer" containerID="4dae7561a5266c1bfa848994bbf2ed293d84f5e77cbf4543717ac62eadfeb023" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.819212 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j8vkp/must-gather-ztcs8" Nov 25 12:09:33 crc kubenswrapper[4696]: I1125 12:09:33.839430 4696 scope.go:117] "RemoveContainer" containerID="4c1efc41bbf359ff7a4b7763d77266714366da4d7944dc489985f892c75c079f" Nov 25 12:09:34 crc kubenswrapper[4696]: I1125 12:09:34.052956 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" path="/var/lib/kubelet/pods/16cc0183-a87c-4fab-92cb-35f2e7570194/volumes" Nov 25 12:09:38 crc kubenswrapper[4696]: I1125 12:09:38.610206 4696 scope.go:117] "RemoveContainer" containerID="44c80f26246c538d0ed71345d5f80163ed11682fb7476ada5db84557b10e9e47" Nov 25 12:09:38 crc kubenswrapper[4696]: I1125 12:09:38.633318 4696 scope.go:117] "RemoveContainer" containerID="1075a45d422742176dc2b49f0a650153ec4ddc11f6f8c968b0ae12ee52d72266" Nov 25 12:09:47 crc kubenswrapper[4696]: I1125 12:09:47.042398 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:09:47 crc kubenswrapper[4696]: E1125 12:09:47.043087 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2mdzn_openshift-machine-config-operator(ac9983ff-4a9b-456b-8a9c-7e5857aad5d2)\"" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" Nov 25 12:10:01 crc kubenswrapper[4696]: I1125 12:10:01.042870 4696 scope.go:117] "RemoveContainer" containerID="0d495efe9e7c681e5006ee4af9c39117eada86ba9a4955b850848fa670883dc1" Nov 25 12:10:02 crc kubenswrapper[4696]: I1125 12:10:02.113875 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" event={"ID":"ac9983ff-4a9b-456b-8a9c-7e5857aad5d2","Type":"ContainerStarted","Data":"33ea71c4080723cad2f6f510a57677958655f1f93bcec0cf6b5d36f9166c4cff"} Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.880265 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s2l7v"] Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881156 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="registry-server" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881169 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="registry-server" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881187 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="copy" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881195 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="copy" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881211 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="extract-content" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881219 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="extract-content" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881237 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="extract-utilities" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881246 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="extract-utilities" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881270 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="extract-utilities" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881277 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="extract-utilities" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881294 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881301 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881325 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="extract-content" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881332 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="extract-content" Nov 25 12:10:43 crc kubenswrapper[4696]: E1125 12:10:43.881340 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="gather" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881347 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="gather" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881540 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d11d67-e798-4b94-be9f-d7bb916065a0" containerName="registry-server" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881558 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="copy" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881579 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbc8d0c-14cc-41e9-bb7f-18aefbc8071c" containerName="registry-server" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.881587 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="16cc0183-a87c-4fab-92cb-35f2e7570194" containerName="gather" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.883117 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.905075 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2l7v"] Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.921122 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-catalog-content\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.921481 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlzw4\" (UniqueName: \"kubernetes.io/projected/796d1ae7-069e-418a-ab8c-7328cb76267c-kube-api-access-hlzw4\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:43 crc kubenswrapper[4696]: I1125 12:10:43.921653 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-utilities\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.023051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlzw4\" (UniqueName: \"kubernetes.io/projected/796d1ae7-069e-418a-ab8c-7328cb76267c-kube-api-access-hlzw4\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.023382 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-utilities\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.023593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-catalog-content\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.023897 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-utilities\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.024085 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-catalog-content\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.060508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlzw4\" (UniqueName: \"kubernetes.io/projected/796d1ae7-069e-418a-ab8c-7328cb76267c-kube-api-access-hlzw4\") pod \"redhat-marketplace-s2l7v\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.217879 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.481098 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wfj6l"] Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.486834 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.494072 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfj6l"] Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.635605 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvg2g\" (UniqueName: \"kubernetes.io/projected/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-kube-api-access-fvg2g\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.635724 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-catalog-content\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.635793 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-utilities\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.686158 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2l7v"] Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.785525 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvg2g\" (UniqueName: \"kubernetes.io/projected/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-kube-api-access-fvg2g\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.785616 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-catalog-content\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.785710 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-utilities\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.786348 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-utilities\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.787894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-catalog-content\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:44 crc kubenswrapper[4696]: I1125 12:10:44.805459 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvg2g\" (UniqueName: \"kubernetes.io/projected/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-kube-api-access-fvg2g\") pod \"certified-operators-wfj6l\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:45 crc kubenswrapper[4696]: I1125 12:10:45.163275 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:45 crc kubenswrapper[4696]: I1125 12:10:45.606488 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerStarted","Data":"2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50"} Nov 25 12:10:45 crc kubenswrapper[4696]: I1125 12:10:45.606729 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerStarted","Data":"03f9638047e1271118f00516c054d8a9452971bba49fdc3ab4dfbfbca856b8cd"} Nov 25 12:10:45 crc kubenswrapper[4696]: I1125 12:10:45.729303 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfj6l"] Nov 25 12:10:46 crc kubenswrapper[4696]: I1125 12:10:46.617521 4696 generic.go:334] "Generic (PLEG): container finished" podID="796d1ae7-069e-418a-ab8c-7328cb76267c" containerID="2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50" exitCode=0 Nov 25 12:10:46 crc kubenswrapper[4696]: I1125 12:10:46.617565 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerDied","Data":"2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50"} Nov 25 12:10:46 crc kubenswrapper[4696]: I1125 12:10:46.619692 4696 generic.go:334] "Generic (PLEG): container finished" podID="8d1560e4-d690-4aee-b70b-19ccdc38b7ab" containerID="2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20" exitCode=0 Nov 25 12:10:46 crc kubenswrapper[4696]: I1125 12:10:46.619728 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerDied","Data":"2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20"} Nov 25 12:10:46 crc kubenswrapper[4696]: I1125 12:10:46.619751 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerStarted","Data":"b351669b1eb41f65402318fd65070f148b8e68b3df8349ed82a9dbf5e16ecb85"} Nov 25 12:10:46 crc kubenswrapper[4696]: I1125 12:10:46.621334 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:10:47 crc kubenswrapper[4696]: I1125 12:10:47.634133 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerStarted","Data":"f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3"} Nov 25 12:10:47 crc kubenswrapper[4696]: I1125 12:10:47.639720 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerStarted","Data":"1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40"} Nov 25 12:10:48 crc kubenswrapper[4696]: I1125 12:10:48.652109 4696 generic.go:334] "Generic (PLEG): container finished" podID="796d1ae7-069e-418a-ab8c-7328cb76267c" containerID="f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3" exitCode=0 Nov 25 12:10:48 crc kubenswrapper[4696]: I1125 12:10:48.652236 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerDied","Data":"f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3"} Nov 25 12:10:49 crc kubenswrapper[4696]: I1125 12:10:49.667110 4696 generic.go:334] "Generic (PLEG): container finished" podID="8d1560e4-d690-4aee-b70b-19ccdc38b7ab" containerID="1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40" exitCode=0 Nov 25 12:10:49 crc kubenswrapper[4696]: I1125 12:10:49.668232 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerDied","Data":"1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40"} Nov 25 12:10:49 crc kubenswrapper[4696]: I1125 12:10:49.675405 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerStarted","Data":"78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8"} Nov 25 12:10:49 crc kubenswrapper[4696]: I1125 12:10:49.785021 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s2l7v" podStartSLOduration=4.398644975 podStartE2EDuration="6.784996585s" podCreationTimestamp="2025-11-25 12:10:43 +0000 UTC" firstStartedPulling="2025-11-25 12:10:46.621059193 +0000 UTC m=+5643.434675800" lastFinishedPulling="2025-11-25 12:10:49.007410823 +0000 UTC m=+5645.821027410" observedRunningTime="2025-11-25 12:10:49.761153723 +0000 UTC m=+5646.574770320" watchObservedRunningTime="2025-11-25 12:10:49.784996585 +0000 UTC m=+5646.598613172" Nov 25 12:10:50 crc kubenswrapper[4696]: I1125 12:10:50.691590 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerStarted","Data":"661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6"} Nov 25 12:10:50 crc kubenswrapper[4696]: I1125 12:10:50.719574 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wfj6l" podStartSLOduration=3.115369538 podStartE2EDuration="6.719552897s" podCreationTimestamp="2025-11-25 12:10:44 +0000 UTC" firstStartedPulling="2025-11-25 12:10:46.621243529 +0000 UTC m=+5643.434860116" lastFinishedPulling="2025-11-25 12:10:50.225426888 +0000 UTC m=+5647.039043475" observedRunningTime="2025-11-25 12:10:50.710717408 +0000 UTC m=+5647.524334005" watchObservedRunningTime="2025-11-25 12:10:50.719552897 +0000 UTC m=+5647.533169494" Nov 25 12:10:54 crc kubenswrapper[4696]: I1125 12:10:54.217939 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:54 crc kubenswrapper[4696]: I1125 12:10:54.218435 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:54 crc kubenswrapper[4696]: I1125 12:10:54.265422 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:54 crc kubenswrapper[4696]: I1125 12:10:54.783967 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:55 crc kubenswrapper[4696]: I1125 12:10:55.164929 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:55 crc kubenswrapper[4696]: I1125 12:10:55.165000 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:55 crc kubenswrapper[4696]: I1125 12:10:55.250359 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:55 crc kubenswrapper[4696]: I1125 12:10:55.464500 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2l7v"] Nov 25 12:10:55 crc kubenswrapper[4696]: I1125 12:10:55.793294 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:56 crc kubenswrapper[4696]: I1125 12:10:56.751579 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s2l7v" podUID="796d1ae7-069e-418a-ab8c-7328cb76267c" containerName="registry-server" containerID="cri-o://78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8" gracePeriod=2 Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.262871 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.376885 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-utilities\") pod \"796d1ae7-069e-418a-ab8c-7328cb76267c\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.377059 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlzw4\" (UniqueName: \"kubernetes.io/projected/796d1ae7-069e-418a-ab8c-7328cb76267c-kube-api-access-hlzw4\") pod \"796d1ae7-069e-418a-ab8c-7328cb76267c\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.377246 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-catalog-content\") pod \"796d1ae7-069e-418a-ab8c-7328cb76267c\" (UID: \"796d1ae7-069e-418a-ab8c-7328cb76267c\") " Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.377999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-utilities" (OuterVolumeSpecName: "utilities") pod "796d1ae7-069e-418a-ab8c-7328cb76267c" (UID: "796d1ae7-069e-418a-ab8c-7328cb76267c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.382332 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/796d1ae7-069e-418a-ab8c-7328cb76267c-kube-api-access-hlzw4" (OuterVolumeSpecName: "kube-api-access-hlzw4") pod "796d1ae7-069e-418a-ab8c-7328cb76267c" (UID: "796d1ae7-069e-418a-ab8c-7328cb76267c"). InnerVolumeSpecName "kube-api-access-hlzw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.396986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "796d1ae7-069e-418a-ab8c-7328cb76267c" (UID: "796d1ae7-069e-418a-ab8c-7328cb76267c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.479966 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.480017 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlzw4\" (UniqueName: \"kubernetes.io/projected/796d1ae7-069e-418a-ab8c-7328cb76267c-kube-api-access-hlzw4\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.480031 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796d1ae7-069e-418a-ab8c-7328cb76267c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.666032 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfj6l"] Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.765583 4696 generic.go:334] "Generic (PLEG): container finished" podID="796d1ae7-069e-418a-ab8c-7328cb76267c" containerID="78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8" exitCode=0 Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.765632 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerDied","Data":"78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8"} Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.766839 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2l7v" event={"ID":"796d1ae7-069e-418a-ab8c-7328cb76267c","Type":"ContainerDied","Data":"03f9638047e1271118f00516c054d8a9452971bba49fdc3ab4dfbfbca856b8cd"} Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.765673 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2l7v" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.766902 4696 scope.go:117] "RemoveContainer" containerID="78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.767907 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wfj6l" podUID="8d1560e4-d690-4aee-b70b-19ccdc38b7ab" containerName="registry-server" containerID="cri-o://661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6" gracePeriod=2 Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.803167 4696 scope.go:117] "RemoveContainer" containerID="f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.836603 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2l7v"] Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.843158 4696 scope.go:117] "RemoveContainer" containerID="2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.854598 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2l7v"] Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.901095 4696 scope.go:117] "RemoveContainer" containerID="78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8" Nov 25 12:10:57 crc kubenswrapper[4696]: E1125 12:10:57.901619 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8\": container with ID starting with 78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8 not found: ID does not exist" containerID="78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.901647 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8"} err="failed to get container status \"78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8\": rpc error: code = NotFound desc = could not find container \"78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8\": container with ID starting with 78b9f38e8d275548921cce8bff1d4345906a09327a755f5f3f6b43461ec5f7a8 not found: ID does not exist" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.901694 4696 scope.go:117] "RemoveContainer" containerID="f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3" Nov 25 12:10:57 crc kubenswrapper[4696]: E1125 12:10:57.901943 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3\": container with ID starting with f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3 not found: ID does not exist" containerID="f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.901966 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3"} err="failed to get container status \"f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3\": rpc error: code = NotFound desc = could not find container \"f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3\": container with ID starting with f2ea6cab28a66986ea729403e3befe94f613e873656c37e83c7564f62c9129f3 not found: ID does not exist" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.901979 4696 scope.go:117] "RemoveContainer" containerID="2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50" Nov 25 12:10:57 crc kubenswrapper[4696]: E1125 12:10:57.902235 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50\": container with ID starting with 2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50 not found: ID does not exist" containerID="2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50" Nov 25 12:10:57 crc kubenswrapper[4696]: I1125 12:10:57.902254 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50"} err="failed to get container status \"2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50\": rpc error: code = NotFound desc = could not find container \"2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50\": container with ID starting with 2c1f82320024a45d235165cc70f83051f519f5e8c997f65a77536c7de66f7c50 not found: ID does not exist" Nov 25 12:10:57 crc kubenswrapper[4696]: E1125 12:10:57.955344 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod796d1ae7_069e_418a_ab8c_7328cb76267c.slice\": RecentStats: unable to find data in memory cache]" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.054604 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="796d1ae7-069e-418a-ab8c-7328cb76267c" path="/var/lib/kubelet/pods/796d1ae7-069e-418a-ab8c-7328cb76267c/volumes" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.710718 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.781780 4696 generic.go:334] "Generic (PLEG): container finished" podID="8d1560e4-d690-4aee-b70b-19ccdc38b7ab" containerID="661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6" exitCode=0 Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.781843 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerDied","Data":"661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6"} Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.782133 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfj6l" event={"ID":"8d1560e4-d690-4aee-b70b-19ccdc38b7ab","Type":"ContainerDied","Data":"b351669b1eb41f65402318fd65070f148b8e68b3df8349ed82a9dbf5e16ecb85"} Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.782160 4696 scope.go:117] "RemoveContainer" containerID="661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.781875 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfj6l" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.800043 4696 scope.go:117] "RemoveContainer" containerID="1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.813658 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-catalog-content\") pod \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.813723 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-utilities\") pod \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.813755 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvg2g\" (UniqueName: \"kubernetes.io/projected/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-kube-api-access-fvg2g\") pod \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\" (UID: \"8d1560e4-d690-4aee-b70b-19ccdc38b7ab\") " Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.815776 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-utilities" (OuterVolumeSpecName: "utilities") pod "8d1560e4-d690-4aee-b70b-19ccdc38b7ab" (UID: "8d1560e4-d690-4aee-b70b-19ccdc38b7ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.820841 4696 scope.go:117] "RemoveContainer" containerID="2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.823272 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-kube-api-access-fvg2g" (OuterVolumeSpecName: "kube-api-access-fvg2g") pod "8d1560e4-d690-4aee-b70b-19ccdc38b7ab" (UID: "8d1560e4-d690-4aee-b70b-19ccdc38b7ab"). InnerVolumeSpecName "kube-api-access-fvg2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.868701 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d1560e4-d690-4aee-b70b-19ccdc38b7ab" (UID: "8d1560e4-d690-4aee-b70b-19ccdc38b7ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.906840 4696 scope.go:117] "RemoveContainer" containerID="661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6" Nov 25 12:10:58 crc kubenswrapper[4696]: E1125 12:10:58.907267 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6\": container with ID starting with 661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6 not found: ID does not exist" containerID="661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.907300 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6"} err="failed to get container status \"661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6\": rpc error: code = NotFound desc = could not find container \"661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6\": container with ID starting with 661382c4b09b9434eff9f73781dcc4b2f69f8655dcf10b25f3f70960c7eb41f6 not found: ID does not exist" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.907322 4696 scope.go:117] "RemoveContainer" containerID="1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40" Nov 25 12:10:58 crc kubenswrapper[4696]: E1125 12:10:58.907563 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40\": container with ID starting with 1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40 not found: ID does not exist" containerID="1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.907589 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40"} err="failed to get container status \"1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40\": rpc error: code = NotFound desc = could not find container \"1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40\": container with ID starting with 1bc097d9b71b83e9c2056fcfb986985c1e8668ebea2007ef0d02bfaa65f84c40 not found: ID does not exist" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.907602 4696 scope.go:117] "RemoveContainer" containerID="2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20" Nov 25 12:10:58 crc kubenswrapper[4696]: E1125 12:10:58.908071 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20\": container with ID starting with 2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20 not found: ID does not exist" containerID="2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.908101 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20"} err="failed to get container status \"2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20\": rpc error: code = NotFound desc = could not find container \"2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20\": container with ID starting with 2c9195abc1b51a89ed80cfe726310eaf86dcdd5a957528b3b1b2187ced460f20 not found: ID does not exist" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.916478 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.916523 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:58 crc kubenswrapper[4696]: I1125 12:10:58.916534 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvg2g\" (UniqueName: \"kubernetes.io/projected/8d1560e4-d690-4aee-b70b-19ccdc38b7ab-kube-api-access-fvg2g\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:59 crc kubenswrapper[4696]: I1125 12:10:59.121203 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfj6l"] Nov 25 12:10:59 crc kubenswrapper[4696]: I1125 12:10:59.135041 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wfj6l"] Nov 25 12:11:00 crc kubenswrapper[4696]: I1125 12:11:00.052351 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1560e4-d690-4aee-b70b-19ccdc38b7ab" path="/var/lib/kubelet/pods/8d1560e4-d690-4aee-b70b-19ccdc38b7ab/volumes" Nov 25 12:12:30 crc kubenswrapper[4696]: I1125 12:12:30.801826 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:12:30 crc kubenswrapper[4696]: I1125 12:12:30.802384 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:13:00 crc kubenswrapper[4696]: I1125 12:13:00.801722 4696 patch_prober.go:28] interesting pod/machine-config-daemon-2mdzn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:13:00 crc kubenswrapper[4696]: I1125 12:13:00.802315 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2mdzn" podUID="ac9983ff-4a9b-456b-8a9c-7e5857aad5d2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111316530024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111316531017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111302663016503 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111302664015454 5ustar corecore